Sorry for asking a very basic question, but my access to textbooks
is rather limited right now, and I'm helping out a friend with a
little medical research project without having much experience in
this field myself.
We want to test for differences in the effect of two drugs on
several variables, some of which have both non-normal
distributions and heterogeneous variances (the target variables
are waiting times, error rates with certain tasks and the like, so
there certainly are some useful transformations which might deal
both with nonnormality and heterogeneiety, but this is another
issue).
Neither the deviations from normality nor the heterogeneiety are
very severe in most cases, so we have calculated both the t-test
(group sizes are equal, by the way) and the Mann-Whitney-U-test,
the latter being a fallback line when the assumptions of the
t-test don't hold. Especially in case both tests come out
differently, we think that further investigation and maybe
transformations seem worthwhile.
Now here is the problem: The mean differences for the t-test are
always in the expected direction, but the 'Mean Ranks' that SPSS
prints out with the U-test are inconsistent with the means. I.e.,
when the t-test shows a significant difference in the expected
direction, the U-test is also significant, BUT the group with the
higher mean also has the higher mean rank. For some very similar
variables (e.g., repetetions of the same tests after a fixed time
- maybe I'm going to ask about repeated measures soon ;-)) the
mean difference shows an almost constant difference between
the two drugs, but again the mean ranks switch place.
I'm not sure that I fully understand what the U-test does, but
could this be due to outliers which affect the means but who loose
there effect when the ranking is done? But if this was the reason
- why do both tests show very similar significance levels? Just
by chance in our data?
Please enlighten me! (References to introductory texts are also
welcome, of course. Textbook access will hopefully improve soon).
Many thanks, MQ
--
--------------------------------------------------------------------
Markus Quandt qua...@wiso.Uni-Koeln.DE
Universitaet zu Koeln
University of Cologne, Germany
>Sorry for asking a very basic question, but my access to textbooks is
>rather limited right now, and I'm helping out a friend with a little
>medical research project without having much experience in this field
>myself.
>We want to test for differences in the effect of two drugs on several
>variables, some of which have both non-normal
>distributions and heterogeneous variances (the target variables are
>waiting times, error rates with certain tasks and the like, so there
>certainly are some useful transformations which might deal both with
>nonnormality and heterogeneiety, but this is another issue).
>Neither the deviations from normality nor the heterogeneiety are very
>severe in most cases, so we have calculated both the t-test (group sizes
>are equal, by the way) and the Mann-Whitney-U-test, the latter being a
>fallback line when the assumptions of the t-test don't hold. Especially
>in case both tests come out
>differently, we think that further investigation and maybe
>transformations seem worthwhile.
>Now here is the problem: The mean differences for the t-test are always
>in the expected direction, but the 'Mean Ranks' that SPSS prints out
>with the U-test are inconsistent with the means. I.e., when the t-test
>shows a significant difference in the expected direction, the U-test is
>also significant, BUT the group with the higher mean also has the higher
>mean rank. For some very similar variables (e.g., repetetions of the
I don't see a problem here; higher means usually have higher mean ranks. A
known error in SPSS Mann-Whitney is that the sign of the printed Z-test is
negative regardless of the dircetion of the difference between groups.
Maybe that is what you mean ?
>same tests after a fixed time - maybe I'm going to ask about repeated
>measures soon ;-)) the mean difference shows an almost constant
>difference between the two drugs, but again the mean ranks switch place.
>I'm not sure that I fully understand what the U-test does, but could this
The Mann-Whitney U statistic counts the number of pairs X and Y (X i group
1, Y in group 2) with X<Y. It is equivalent to the sum of the ranks in
either group, hence the name 'ranksum test'.
>be due to outliers which affect the means but who loose there effect when
>the ranking is done? But if this was the reason - why do both tests show
>very similar significance levels? Just by chance in our data?
When groups are fairly large, say >30, you would expect that. A somewhat
technical reference for rank tests is Lehmann: Nonparametrics. Wiley 1975.
>Please enlighten me! (References to introductory texts are also welcome,
>of course. Textbook access will hopefully improve soon).
>Many thanks, MQ
Hope this is helpful,
--
Arnold Kester
Arnold...@stat.unimaas.nl
From Arnold's response I see that I didn't express the problem very
clearly, so I'll try to reformulate it at the critical places.
(Sorry, the quotation is quite messed up).
Arnold Kester wrote:
>
> In <dyvR14AN...@wiso.Uni-Koeln.DE>, on 04/29/98
> at 12:10 PM, qua...@wiso.Uni-Koeln.DE (Markus Quandt) said:
snip
>
> >Now here is the problem: The mean differences for the t-test are always
> >in the expected direction, but the 'Mean Ranks' that SPSS prints out
> >with the U-test are inconsistent with the means. I.e., when the t-test
> >shows a significant difference in the expected direction, the U-test is
> >also significant, BUT the group with the higher mean also has the higher
> >mean rank.
>
> I don't see a problem here; higher means usually have higher mean ranks.
I assumed that an ascending ranking is applied: a high value of the
measurement would then have led to a *numerically* low rank, i.e. the
highest measurement value gets rank place '1', the next '2' and so on.
Arnold, thanks for clearing this up, but this is exactly where my
problem is: Sometimes the group with the higher mean has the higher
rank, sometimes it has the lower rank. So, my question is: What are the
exceptions to 'usually'?
> the mean difference shows an almost constant
> difference between the two drugs, but again the mean ranks switch place.
>
[Could this...]
> >be due to outliers which affect the means but who loose there effect when
> >the ranking is done? But if this was the reason - why do both tests show
> >very similar significance levels? Just by chance in our data?
Since the above failed so badly in getting the problem straight, I'll
try again and give an example. (This is not the only case where the
phenomenon occurs).
Suppose you have repeated measurements of the effects of each of our two
drugs at four points in time. The means and mean ranks of the test come
out as follows:
meas. drug1 drug2 t-test sign. drug1 drug2 U-test sign.
1 -1.34 -6.84 yes 63.5 37.5 yes
2 3.56 -.2 yes 45.5 55.5 yes
3 5.8 2.26 yes 49.1 51.9 no
4 6.674 4.38 yes 44.1 56.9 yes
You see that the order of the mean ranks is switching between the two
drugs, while the order of the means is not. How does it happen, and what
does it imply for the interpretation?
Regards, MQ
: From Arnold's response I see that I didn't express the problem very
: clearly, so I'll try to reformulate it at the critical places.
<< snip >>
: Since the above failed so badly in getting the problem straight, I'll
: try again and give an example. (This is not the only case where the
: phenomenon occurs).
<< snip >>
I think you better look at histograms of your groups, or box-and-
whisker plots.
The gross assumption of the rank-test is that the distributions
are the same general shape - if that were the case, then it
would be impossible for the t-test to reject in the OPPOSITE
direction, unless you are (improperly) ignoring some outliers.
If one group is better on the t-test and
the other has the rank-order superiority, then
it is ALMOST BOUND TO BE TRUE that NEITHER test
is legitimate, i.e., wholly believable. The assumptions
are not so very different. So you may have to report both tests,
or else, make a choice : What assumptions do you want to throw away?
--
Rich Ulrich, biostatistician wpi...@pitt.edu
http://www.pitt.edu/~wpilib/index.html Univ. of Pittsburgh