I'm adding this to a new thread to keep 'should disagreeing with the
community ID make an observation "casual"?' on topic,
but I thought the exchange between Wouter et al about disagreements
interesting. Just to provide some numbers to Jakob's claims, there are
3,598 obs that both Jakob and Wouter have ID'd and it looks like Jakob
and Wouter's IDs are 'maverick' to each other (this is a word we've
been using to describe IDs that are not descendants or ancestors of
one another) only 2 times:
http://www.inaturalist.org/observations/812934
http://www.inaturalist.org/observations/5101903
So thats a 99.94% agreement which is a disagreement rate of < 1 in 1,000
I'd be very interested in trying to quantify the average 'error rate'
in iNat obs IDs. The biggest challenge towards doing this is finding a
metric of 'truth' that we can use to filter obs with IDs we are very
confident in for analyses. As has been pointed out, even 'experts' on
the site seem to have error rates of at least 1%. This is probably do
to a combination of real disagreement, accidentally clicking on the
wrong thing, confusion about the subject of an observation (the bee or
the flower), or taxonomic issues (do you mean the genus sensu-lato or
sensu-stricto?)
I've been playing with an approach of finding 'truth' that looks at
sample of 20k observations that have been ID'd by at least 3 IDers
like jakob with lots of ID's and low empirical error rates (ie novel
IDs that the community rejected over novel IDs the community embraced
- Jakob's empirical 'error' rate using this approach is 3% which is
the same as mine). If we call truth the common ancestor of these ID's
than you can assume that these obs are mis-ID'd the probability of all
3 people wrong ~ (1-0.03*0.03*0.03)*100 << 0.01 = 1 in 1,000
Assuming this 'truth' set is a somewhat representative sample, we can
use it to try to predict what kinds of observations tend to be
mis-ID'd and see what factors contribute to more accurate IDs (e.g.
things like rank, taxonomic clade, location, reputation of the IDer).
If we were to implement some sort of more sophisticated test for what
obs we think are 'research grade' over what we're doing now, I'd
personally like it to be informed by analyses along these lines.
-Scott
--
--------------------------------------------------
Scott R. Loarie, Ph.D.
Co-director, iNaturalist.org
California Academy of Sciences
55 Music Concourse Dr
San Francisco, CA 94118
--------------------------------------------------