Thanks for the feedback - we definitely share your concerns and are
monitoring how new suggestions tool is monitoring behavior on the site
so we can tweak things as necessary. A few things:
1) We're tracking every ID that was made using the suggestions tool.
But this does get complicated, because using the ID tool doesn't
necessarily mean the user didn't also know the ID. I personally find
it useful as a kind of visual autocomplete that saves me time not
having to type in names for species I know.
2) We are currently presenting them as 'suggestions' rather than
automatically suggesting an ID. Some quick stats: Based on a test
sample of 50k photos representative of everything thats been posted to
iNat (so that includes some common things, but also some very rare
things) the suggestions tool places the correct species is in the top
10 results 78% of the time, in the top 2 results 66% of the time, and
in the top spot 57% of the time.
But we also tried to encourage people not to go straight to the top
suggestion by adding in the coarser ranked 'recommendation' at the
top. We are able to produce a common ancestor 77% of the time
(otherwise it says we're not confident enough to recommend something).
If it does produce a recommendation, its right 93% of the time. A key
goal is to bring all these stats up, but I'd like in particular the
recommendation accuracy to be higher since thats the choice we're
actually encouraging you to tap on
3) In parallel to this computer vision work, we're also conducting an
analysis of the quality of the identifications on iNat you can read
more and help here:
https://www.inaturalist.org/pages/identification_quality_experiment
We're a little behind on this (having gotten bogged down in computer
vision) but have more time for it now. The main goals of the study are
to see if we can assign a quantitative accuracy threshold to 'research
grade' - right now it means >2/3 people agree. We're working on a
model that incorporates more data about an observation such as each
IDer's past behavior (earned reputation) to come up with a
quantitative predicted estimate that an observation is accurately
ID'd. Once we have this we'll be able to adjust Research Grade so that
its some threshold (ie obs with >99% predicted accuracy).
I think this would solve issues associated with trusting the AI
because we'd be able to be explicit about some level of accuracy
associated with an observation ID (whether it comes from a completely
anonymous user using the suggestions tool, or some mix of that and the
opinions of community) so folks could judge for themselves what level
of accuracy they're comfortable with. Agreeing on a threshold for
Research Grade might be tricky though - should it include errors of 1
in 100, 1 in 1000 etc?
Scott
> --
> You received this message because you are subscribed to the Google Groups
> "iNaturalist" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to
inaturalist...@googlegroups.com.
> To post to this group, send email to
inatu...@googlegroups.com.
> Visit this group at
https://groups.google.com/group/inaturalist.
> For more options, visit
https://groups.google.com/d/optout.
--
--------------------------------------------------
Scott R. Loarie, Ph.D.
Co-director, iNaturalist.org
California Academy of Sciences
55 Music Concourse Dr
San Francisco, CA 94118
--------------------------------------------------