I would be interested to hear what other people suggest.
Francesca
Quoting Liz <lizh...@hotmail.com>:
Francesca Chappell
Medical Statistician
University of Edinburgh
Bramwell Dott Building
Western General Hospital
Crewe Road
Edinburgh EH4 2XU
Tel 0131 537 3585
Fax 0131 332 5150
The University of Edinburgh is a charitable body, registered in Scotland, with
registration number SC005336
--
The University of Edinburgh is a charitable body, registered in
Scotland, with registration number SC005336.
Francesca
Quoting Ray Koopman <koo...@sfu.ca>:
Best,
Martin
I've not been following this thread carefully, excuse me if this has been covered.
If you have specificity of 100 and sensitivity much lower (20-50%) then there is something seriously wrong. Isn't this equivalent to jsut telling everyone they have the disease? So, I think you need to adjust your model.
Peter
Peter L. Flom, PhD
Statistical Consultant
www DOT peterflom DOT com
But to digress again and if you're not constrained to use the
chains-of-LRs route, but some people estimate the sensitivity and
specificity of test combinations, where you use Boolean AND or OR to
combine the individual test results and you would avoid the zero
divisor problem. For example, you could use the OR rule (any test is
positive means the diagnosis is positive) to increase sensitivity,
though of course specificity would go down. Haven't read the Rudwaleit
et al paper so can't comment on that.
Francesca
Maybe, maybe not. Unless there is some huge problem with giving people a false positive, then I think you may need to adjust the point at which you diagnose inflammatory arthritis. What if you got 90% specificity, would you get much better sensitivity?
Usually, I think, guaranteeing no false negatives would increase false positives by a lot....
But I overstated it in my previous message.
Its difficult to draw probability trees, but if you set out your test
series as branches, and work out some rules about how many tests have
to be +ve before they are classed as true +ve's and -ve before they
are classified as true -ves. Then its quite easy to compute your
sensitivity and specificity of the combination of tests you want to
use.
Adrian
2008/6/19 Liz <lizh...@hotmail.com>:
Specificty, as I understand it, is a measure of the probability
that you will not get a positive result, when the condition being
tested for is not present -- a kind of reassurance that a positive
result is indeed specific to that condition: but this interpretation
implicitly has a "reference level" based on the prevalence.
Secondly, now that I look back again throught this thread, I cannot
see that Liz has stated what the sizes of the two groups are.
Whatever the size of the "non-inflammatory" group, an observed 100%
of "test negative" still leads to uncertainty of what the probability
of "positive given not present" really is. Say we assess this as
a 95% confidence interval. Let N be the size of the "non-inflammatory"
group. Then:
N = 50: CI = (0,0.0582)
N = 100: CI = (0,0.0295)
N = 250: CI = (0,0.0119)
N = 500: CI = (0,0.00598)
N = 1000: CI = (0,0.00299)
So you need a big group before you can be reasonably sure that the
probability is quite small; and you are never sure that it is zero.
In order to avoid unrealistic numerical results, one should try to
take account if uncertaqinties of this kind.
At Date: Wed, 18 Jun 2008 06:20:16 -0700 (PDT), Liz stated:
I'm following the methods of Rudwaleit et al.
(Ann Rheum Dis 2004;63;535-543) in which they
calculate the probability of the presence of the
disease in case of a positive test as
100/[1 + ((100-pretestprob)/pretestprob)*
((100-specificity)/sensitivity)].
(I've added the closing "]" in the above). This formula is
just Bayes' theorem for Prob(present given positive-test),
using "pretestprob" to mean "prevalence", i.e. the prior
probability that the condition is present.
If, now (as in Liz'z statement of her problem), the specificity
is 100%, then the factor (100-specificity)/sensitivity) is 0
(provided sensitivity > 0, which Liz states is the case).
Hence the result of that formula is 100/1 = 100%: if specificity
is 100%, then the probability of presence given positive test is
100%, as one would expect; and so no problems arise. So far so good.
Liz goes on to state:
The probability of the presence of disease in the
case of a negative test inverts the last part of that
equation (specificity/100-sensitivity) so if either
sensitivity or specificity is 100 we hit a snag.
This may be where confusion could arise. If I work out (from the
Bayes' theorem direction) what the probability of presence is when
the test is negative, I get (rephrasing it in the above terms):
100/[1 + (specificity/(100 - sensitivity))*
((100 - pretestprob)/pretestprob)]
so, if the specificity is 100% again, then the result still makes
sense (no "snag") provided the sensitivity is not 100%. Liz
states that the sensitivity is 20-50%, so no problem. Even if the
sensitivity were 100%, all that results (mathematically speaking)
is 100/(1 + infinity) = 0. But that's OK too, since 100% sensitivity
means that every true positive gives a positive test, so a negative
test must mean that the disease is not present -- so Prob=0.
So still no problem! Though, depending on how the calculation of
such a formula is implemented in software or programmed in code,
one may get an uninformtaive answer. However, that is dependent
on the software. If, for instance, I do it in R, then (adopting
pretestprob=17% arbitrarily):
spec=100 ; sens=100; pretestprob=17
100/(1 + (spec/(100 - sens))*((100 - pretestprob)/pretestprob))
0
as expected. And this is because R does the right thing with Infinity:
1/0
Inf
1/Inf
0
I don't know if this helps -- maybe there's something more we need
to be told about where the real problem is arising -- but I hope
it helps a bit!
Ted.
On 19-Jun-08 10:39:10, Liz wrote:
>
> Sorry Peter I'm not sure I understand you. I'm looking at predictors
> of inflammatory arthritis. All of the patients who are in the 'non-
> inflammatory' group are test negative (hence specificity=100%),
> whereas between 20% and 50% of those in the 'inflammatory' group are
> test positive, depending on the test in question (sensitivity
> 20%-50%). Is there a problem with my model?
>
>> I've not been following this thread carefully, excuse me if this has
>> been covered.
>>
>> If you have specificity of 100 and sensitivity much lower (20-50%)
>> then there is something seriously wrong. _Isn't this equivalent to
>> jsut telling everyone they have the disease? _So, I think you need to
>> adjust your model.
>>
>> Peter
>>
>> Peter L. Flom, PhD
>> Statistical Consultant
>> www DOT peterflom DOT com
--------------------------------------------------------------------
E-Mail: (Ted Harding) <Ted.H...@manchester.ac.uk>
Fax-to-email: +44 (0)870 094 0861
Date: 19-Jun-08 Time: 13:43:05
------------------------------ XFMail ------------------------------
On 19-Jun-08 12:43:11, Ted Harding wrote:
> [...]
> At Date: Wed, 18 Jun 2008 06:20:16 -0700 (PDT), Liz stated:
>
> I'm following the methods of Rudwaleit et al.
> (Ann Rheum Dis 2004;63;535-543) in which they
> calculate the probability of the presence of the
> disease in case of a positive test as
> 100/[1 + ((100-pretestprob)/pretestprob)*
> ((100-specificity)/sensitivity)].
>
> (I've added the closing "]" in the above). This formula is
> just Bayes' theorem for Prob(present given positive-test),
> using "pretestprob" to mean "prevalence", i.e. the prior
> probability that the condition is present.
This formula can equivalently be re-written (and will then
be in the form arising directly from Bayes' theorem):
The probability of the presence of the disease in case
of a positive test is
sensitivity*pretestprob/
[sensitivity*pretestprob + (100-specificity)*(100-pretestprob)]
If you use it in that form, you will NEVER have any computational
problems! So why adopt a form which could (in certain cases) be
problematic?
Best wishes to all,
Ted.
--------------------------------------------------------------------
E-Mail: (Ted Harding) <Ted.H...@manchester.ac.uk>
Fax-to-email: +44 (0)870 094 0861
Date: 19-Jun-08 Time: 14:01:38
------------------------------ XFMail ------------------------------
For abbreviation, sensitivity is "sens", specificity is "spec",
and pretestprob is "prev" (short for "prevalence").
Then:
Prob(disease given test+) = Prob(disease AND test+)/Prob(test+)
= (Prob(test+ GIVEN disease)*Prob(disease))/Prob(test+)
= (sens*prev)/Prob(test+)
= (sens*prev)/
[Prob(test+ GIVEN disease)*Prob(disease) +
Prob(test+ GIVEN (not disease))*Prob(not disease)]
= (sens*prev)/[sens*prev + (100-spec)*(100-prev)]
QED -- where probabilities are expressed at percentages, of course.
So the form that does not fall over in certain cases is the one
that directly derives from the Bayes approach.
Ted.
On 19-Jun-08 15:01:38, Liz wrote:
>
> Ah!!! Thanks Ted - I was just following the methods exactly as set out
> in the paper, like a statistical sheep. I have no idea why the formula
> was presented like that in the text if it stems from an existing
> theorem that obviously works rather more reliably. Not that hot on
> Bayes - obviously something I need to work on (a revision of basic
> balancing of equations evidently wouldn't hurt either).
> As for the sample size I absolutely take your point - which is why I
> was originally opting to reduce specificity from 100 to 99. This is a
> small exploratory dataset (N=40! often the case with imaging studies
> where MRIs are so expensive) so we'll be saying something along the
> lines of "whilst our findings need to be replicated in a much larger
> study, these particular tests [a,b,c] are the most promising, and if
> the larger study reveals similar sensitivities and specificities to
> those measured here, then in combination these tests would allow us to
> be x% sure of our diagnosis." Rudwaleit et al. reviewed many existing
> studies and combined them to obtain their estimates of sens and spec.
>
> Thanks so much Ted, and everyone else, for your help.
> Liz
>
> On Jun 19, 2:01_pm, (Ted Harding) <Ted.Hard...@manchester.ac.uk>
> wrote:
>> Follow-up: Inj my previous reply I intended to include (but in
>> the heat of the creative moment forgot) the following (the rest
>> suppressed):
>>
>> On 19-Jun-08 12:43:11, Ted Harding wrote:
>>
>> > [...]
>> > At Date: Wed, 18 Jun 2008 06:20:16 -0700 (PDT), Liz stated:
>>
>> > _ I'm following the methods of Rudwaleit et al.
>> > _ (Ann Rheum Dis 2004;63;535-543) in which they
>> > _ calculate the probability of the presence of the
>> > _ disease in case of a positive test as
>> > _ 100/[1 + ((100-pretestprob)/pretestprob)*
>> > _ _ _ _ _ _((100-specificity)/sensitivity)].
>>
>> > (I've added the closing "]" in the above). This formula is
>> > just Bayes' theorem for Prob(present given positive-test),
>> > using "pretestprob" to mean "prevalence", i.e. the prior
>> > probability that the condition is present.
>>
>> This formula can equivalently be re-written (and will then
>> be in the form arising directly from Bayes' theorem):
>>
>> _ The probability of the presence of the disease in case
>> _ of a positive test is
>>
>> _ sensitivity*pretestprob/
>> _ [sensitivity*pretestprob + (100-specificity)*(100-pretestprob)]
>>
>> If you use it in that form, you will NEVER have any computational
>> problems! So why adopt a form which could (in certain cases) be
>> problematic?
>>
>> Best wishes to all,
>> Ted.
>>
>> --------------------------------------------------------------------
>> E-Mail: (Ted Harding) <Ted.Hard...@manchester.ac.uk>
>> Fax-to-email: +44 (0)870 094 0861
>> Date: 19-Jun-08 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ Time: 14:01:38
>> ------------------------------ XFMail ------------------------------
--------------------------------------------------------------------
E-Mail: (Ted Harding) <Ted.H...@manchester.ac.uk>
Fax-to-email: +44 (0)870 094 0861
Date: 19-Jun-08 Time: 16:23:56
------------------------------ XFMail ------------------------------