It would be nice to get a version in for 0.9. I spent a few
unsuccessful days on it earlier this year. But since there are two new
or corrected versions available, it looks like it just needs testing
and a performance comparison.
I won't have time for this, so if anyone volunteers for this, scipy
0.9 should be able to get Fisher's exact.
As an aside:
There is a related ticket for numerical precision problems in the sf
calculations of discrete distributions,
http://projects.scipy.org/scipy/ticket/1218 . However, this requires a
rewrite of the current generic cdf calculation and adding a matching
sf version. The current implementation only works under special
assumptions which luckily are satisfied by all current discrete
distributions. I think, I will find time for this before 0.9.
Josef
_______________________________________________
SciPy-User mailing list
SciPy...@scipy.org
http://mail.scipy.org/mailman/listinfo/scipy-user
I also do not have time for this month.
I briefly looked at the code at pypi link but I do not think it is
good enough for scipy. Also, I do not like when people license code as
'BSD' and there is a comment in cfisher.pyx '# some of this code is
originally from the internet. (thanks)'. Consequently we can not use
that code.
The code with ticket 956 still needs work especially in terms of the
input types and probably the API (like having a function that allows
the user to select either 1 or 2 tailed tests).
Bruce
On Sat, Nov 13, 2010 at 8:50 PM, Â <josef...@gmail.com> wrote:
> http://projects.scipy.org/scipy/ticket/956 and
> http://pypi.python.org/pypi/fisher/ have Fisher's exact
> testimplementations.
>
> It would be nice to get a version in for 0.9. I spent a few
> unsuccessful days on it earlier this year. But since there are two new
> or corrected versions available, it looks like it just needs testing
> and a performance comparison.
>
> I won't have time for this, so if anyone volunteers for this, scipy
> 0.9 should be able to get Fisher's exact.
I briefly looked at the code at pypi link but I do not think it is
good enough for scipy. Also, I do not like when people license code as
'BSD' and there is a comment in cfisher.pyx  '# some of this code is
originally from the internet. (thanks)'. Consequently we can not use
that code.
The code with ticket 956 still needs work especially in terms of the
input types and probably the API (like having a function that allows
the user to select either 1 or 2 tailed tests).
adding 1 tailed tests would be a nice bonus.
I think, we should add them as much as possible. Currently one-sided
versus two-sided is still somewhat inconsistent across functions. I
added one-sided tests to some functions.
Tests based on symmetric distributions (t or normal) like the t-tests
don't necessarily need both because the one sided test can essentially
be recovered from the two sided test, half or double the pvalue.
I added a comment to the ticket, fisher3 looks good except for the
python 2.4 incompatibility.
Thanks Ralf for taking care of this,
Josef
>
> Cheers,
> Ralf
1) It just can not use np.asarray() without checking the input first. This is particularly bad for masked arrays.
2) There are no dimension checking because, as I understand it, this can only handle a '2 by 2' table. I do not know enough for general 'r by c' tables or the 1-d case either.
3) The odds-ratio should be removed because it is not part of the test. It is actually more general than this test.
4) Variable names such as min and max should not shadow Python functions.
5) Is there a reference to the algorithm implemented? For example, SPSS provides a simple 2 by 2 algorithm:
http://support.spss.com/ProductsExt/SPSS/Documentation/Statistics/algorithms/14.0/app05_sig_fisher_exact_test.pdf
6) Why exactly does the dtype need to int64? That is, is there something wrong with hypergeom function? I just want to understand why the precision change is required because the input should enter with sufficient precision.
There is some discussion in the ticket about more than 2by2,
additions would be nice (and there are some examples on the matlab
fileexchange), but 2by2 is the most common case and has an unambiguous
definition.
>
>>
>> 3) The odds-ratio should be removed because it is not part of the test. It
>> is actually more general than this test.
>>
> Don't feel strongly about this either way. It comes almost for free, and R
> seems to do the same.
same here, it's kind of traditional to return two things, but in this
case the odds ratio is not the test statistic, but I don't see that it
hurts either
>
>> 4) Variable names such as min and max should not shadow Python functions.
>
> Yes, Josef noted this already, will change.
>>
>> 5) Is there a reference to the algorithm implemented? For example, SPSS
>> provides a simple 2 by 2 algorithm:
>>
>> http://support.spss.com/ProductsExt/SPSS/Documentation/Statistics/algorithms/14.0/app05_sig_fisher_exact_test.pdf
>
> Not supplied, will ask on the ticket and include it.
I thought, I saw it somewhere, but don't find the reference anymore,
some kind of bisection algorithm, but having a reference would be
good.
Whatever the algorithm is, it's fast, even for larger values.
>>
>> 6) Why exactly does the dtype need to int64? That is, is there something
>> wrong with hypergeom function? I just want to understand why the precision
>> change is required because the input should enter with sufficient precision.
>>
> This test:
> fisher_exact(np.array([[18000, 80000], [20000, 90000]]))
> becomes much slower and gives an overflow warning with int32. int32 is just
> not enough. This is just an implementation detail and does not in any way
> limit the accepted inputs, so I don't see a problem here.
for large numbers like this the chisquare test should give almost the
same results, it looks pretty "asymptotic" to me. (the usual
recommendation for the chisquare is more than 5 expected observations
in each cell)
I think the precision is required for some edge cases when
probabilities get very small. The main failing case, I was fighting
with for several days last winter, and didn't manage to fix had a zero
at the first position. I didn't think about increasing the precision.
>
> Don't know what the behavior should be if a user passes in floats though?
> Just convert to int like now, or raise a warning?
I wouldn't do any type checking, and checking that floats are almost
integers doesn't sound really necessary either, unless or until users
complain. The standard usage should be pretty clear for contingency
tables with count data.
Josef
>
> Cheers,
> Ralf
It looks good to me. I think you can commit whenever you want. We can
add the reference also later through the doceditor.
Thanks for finding the hypergeom imprecision example. We can use it
for the sf ticket.
I think, from a statistical viewpoint this imprecision is pretty
irrelevant, whether a pvalue is 1e-9 or 1e-8, wouldn't change the
conclusion about strongly rejecting the null hypothesis.
I usually use almost_equal instead of approx_equal which often
wouldn't even fail on differences like this. Many statistical packages
report only 3 decimals by default because everything else only gives a
false sense of precision given the randomness of the problem.
However, independently of this, discrete cdf and sf need a workover,
but maybe not so urgently if these are the only problems.
Sorry but I don't agree. But I said I do not have time to address this
and I really do not like adding the code as it is.
Bruce
Sorry but I don't agree. But I said I do not have time to address this
and I really do not like adding the code as it is.
I know and the reason for my negativity is that this commit goes
against what I had proposed to provide single stats functions that
handle the various ndarray types not just the 'standard array. Also it
lacks the flexibility to handle general R by C cases which are very
common. But that requires time to find how to do those cases. The
error is that there is no dimensionality check .
I find it shocking that a statistical test returns a 'odds ratio' that
has nothing to do with the actual test nor with any of the other
related statistical tests like chisquare. If you accept that then we
must immediately add that odds ratio to ALL statistical tests.
In this case, I see no reason to support any other array types. The
function asks for 4 numbers and returns 2 numbers, no nans or missing
values are allowed. (There is nothing to calculate if one of the 4
values is missing.) I haven't checked carefully this time, but, I
think, any array_like type that allows indexing should by correctly
used.
> Also it
> lacks the flexibility to handle general R by C cases which are very
> common. But that requires time to find how to do those cases. The
> error is that there is no dimensionality check .
Handling more than 2x2 would be a nice extension, but is no reason to
not commit the current version.
It might be good to raise a ValueError if the input is not 2x2. I need
to check whether for example 2x3 wouldn't silently produce an
unintended result.
>
> I find it shocking that a statistical test returns a 'odds ratio' that
> has nothing to do with the actual test nor with any of the other
> related statistical tests like chisquare. If you accept that then we
> must immediately add that odds ratio to ALL statistical tests.
R returns the posterior odds ratio, we return the prior odds-ratio. I
don't care much either way, but the test is (implicitly) on the odds
ratio.
Other tests provide their own statistic, which is not always the
actual statistic used in the calculation of the p-value. If it were
not for backward compatibility, I would add lots of things to ALL
statistical tests, as you can see in some tests in statsmodels.
And did I mention: It's really fast.
Josef
>
> Bruce
> _______________________________________________
> SciPy-User mailing list
> SciPy...@scipy.org
> http://mail.scipy.org/mailman/listinfo/scipy-user
>
Handling more than 2x2 would be a nice extension, but is no reason toOn Sun, Nov 21, 2010 at 2:03 PM, Bruce Southey <bsou...@gmail.com> wrote:
> Also it
> lacks the flexibility to handle general R by C cases which are very
> common. But that requires time to find how to do those cases. The
> error is that there is no dimensionality check .
not commit the current version.
It might be good to raise a ValueError if the input is not 2x2. I need
to check whether for example 2x3 wouldn't silently produce an
unintended result.