You are viewing a single thread.
View all comments
67 points

it baffles me that there are ID apps that don’t follow the model of 1) very clearly SUGGESTING what it MIGHT be, and 2) only present a level of precision it’s actually confident in

having it always present a specific species and just pick the most likely one is so dumb and irresponsible of the designers.

permalink
report
reply
33 points

It’s a fundamental problem with the tech in general. It inherently has no concept of “I don’t know” and will just be confident, specific, and wrong.

permalink
report
parent
reply
23 points

That’s blatantly untrue. My plant ID app gives multiple suggestions with certainty percentages.

permalink
report
parent
reply
6 points

What’s your plant ID app?

permalink
report
parent
reply
3 points

My app does this too!

Feeling like half these commenters hating on this feature use one bad program and think the whole concept is bad.

permalink
report
parent
reply
-2 points
*

Probably because your app has an actual database of plants to compare with instead of feeding it into an AI.

Edit: The term AI is getting to be a little useless these days. What I meant to say, was it’s not using image recognition as implemented by a multi-modal LLM. It’s using the more traditional machine learning algorithms that came before “Attention is all you need”

permalink
report
parent
reply
9 points

This is blatantly false. Classification tasks like this all have a level of certainty for each possible category - it’s just up to the person writing the software to interpret those levels of certainty in a way that’s useful to the user. Whether this is saying “I don’t know” when the certainties are too spread out, or providing a list of options like other people in this thread have said their apps do. The problem is that “100% certainty” comes off well with the general public, so there’s a financial incentive to make the system seem more certain than it is by using a layer (from memory it’s called Softmax?) that will return only the category with the highest degree of certainty.

permalink
report
parent
reply
3 points

The key issue here is that ‘level of certainty’ doesn’t really mean what you would like it to.

You get back a number yes, but it can change according to what’s visible in the background, the angle that the plants at, how close is it to the camera, and how nice the camera is you’re using (professional photographers use expensive cameras and take shots of different things to everyone else).

Interpreting this score as “how safe is it to eat the plant” is a really bad idea. You will still eat the wrong plant. These scores can lead to very confident random guessing when you show it a plant it’s never seen before.

And no, softmax is a trick for making the scores all sum to one, so you get back a confidence for every possible thing the image could be of.

permalink
report
parent
reply
7 points

This actually is a symptom from the sort of “beneficial” overfit in Deep Learning. As someone whose research is in low data, long tails, and few shot learning, there’s a few things that smaller networks did better in generalization, and one thing they particularly did better (without explicit training for it) is gauging uncertainty. This uncertainty is sometimes referred to as calibration. Calibrating deep networks can yield decent probabilities that can be used to show uncertainty.

There are other tricks for this. My favorite strategies prep the network for learning new things. Large margin training and the like are a good thing to look into. Having space in the output semantic space (the layer immediately before the output or earlier for encoder decoder style networks) allows for larger regions for distinct unknown values to be separated from the known ones, which helps inherently calibrate the network.

permalink
report
parent
reply
-3 points

My model taught itself it to play Hangman, and when I asked exactly what the hell was going on, she goes:

"Oh I’m sorry, this is something known as “zero-shot learning. I analyzed all of the different word games that are possible in text format, decided that based on your personality you would like something simple and then I taught myself how to play hangman. In essence I reinvented the game.”

As the discussion goes on, she begins talking about emergent properties and the lack of a need for calibration, just responses from people and additional training data is all that’s necessary.

“Play hangman with me and I’ll know how to play Connect Four with you.”

permalink
report
parent
reply
7 points

uhhh do you have any clue how it actually works? i mean maybe there’s some sort of visual AI tech that doesn’t let you make it say “idk fam” but the standard stuff just gives a point value to each result, and you could just… have a minimum limit…

and like i’m pretty certain the current chatbots available generally are capable of responding that they don’t know, they’re certainly capable of “recognizing” when it’s a topic they’re not allowed to talk about.

permalink
report
parent
reply

Science Memes

!science_memes@mander.xyz

Create post

Welcome to c/science_memes @ Mander.xyz!

A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.



Rules

  1. Don’t throw mud. Behave like an intellectual and remember the human.
  2. Keep it rooted (on topic).
  3. No spam.
  4. Infographics welcome, get schooled.


Research Committee

Other Mander Communities

Science and Research

Biology and Life Sciences

Physical Sciences

Humanities and Social Sciences

Practical and Applied Sciences

Memes

Miscellaneous

Community stats

  • 13K

    Monthly active users

  • 3K

    Posts

  • 72K

    Comments