For image recognition models:
I noticed a difference in labeling and confidence when I test an image within the MLfK platform and then on Scratch. In the screenshots, you'll see the label and confidence for the same image -- rock is the accurate label. Why might this discrepancy be happening, and how can I get these results to be consistent with each other?
Thanks!

