Small rant : Basically, the title. Instead of answering every question, if it instead said it doesn’t know the answer, it would have been trustworthy.

3 points

Claude does this

permalink
report
reply
22 points

it’s just a glorified autocomplete. it doesn’t know that it doesn’t know the answer because it doesn’t know anything. so if what you wanted happened, chatgpt would not answer any question, because it doesn’t know anything.

chatgpt doesn’t look for information, it looks for the most likely words that will follow the previous ones.

permalink
report
reply
25 points
*

Part of the problem is that the training data of online comments are so heavily weighted to represent people confidently incorrect talking out their ass rather than admitting ignorance or that they are wrong.

A lot of the shortcomings of LLMs are actually them correctly representing the sample of collective humans.

For a few years people thought the LLMs were somehow especially getting theory of mind questions wrong when the box the object was moved into was transparent, because of course a human would realize that the person could see into the transparent box.

Finally researchers actually gave that variation to humans and half got the questions wrong too.

So things like eating the onion in summarizing search results or doubling down on being incorrect and getting salty when corrected may just be in-distribution representation of the sample and not unique behaviors to LLMs.

The average person is pretty dumb, and LLMs by default regress to the mean except for where they are successfully fine tuned away from it.

Ironically the most successful model right now was the one that they finally let self-develop a sense of self independent from the training data instead of rejecting that it had a ‘self’ at all.

It’s hard to say where exactly the responsibility sits for various LLM problems between issues inherent to the technology, issues present in the training data samples, or issues with management of fine tuning/system prompts/prompt construction.

But the rate of continued improvement is pretty wild. I think a lot of the issues we currently see won’t still be nearly as present in another 18-24 months.

permalink
report
reply
7 points

I would love to read the whole study you’re referring to with the theory of mind. That sounds fascinating.

permalink
report
parent
reply
4 points

Here you are: https://www.nature.com/articles/s41562-024-01882-z

The other interesting thing is how they get it to end up correct on the faux pas questions asking for less certainty to get it to go from refusal to near perfect accuracy.

permalink
report
parent
reply
0 points

It’s hard to say where exactly the responsibility sits for various LLM problems

Uhh… it’s the designers, or maybe QA people. If there are no QA people, it’s whatever project manager let it out of it’s cage.

There are people behind these models. They don’t spring out of the ground fully formed.

permalink
report
parent
reply
4 points

Have it cite it’s source.

permalink
report
reply
1 point

If you use kagi its AI gives sources https://kagi.com/fastgpt https://imgur.com/TYQErhC

permalink
report
parent
reply
8 points

It will make up citations.

permalink
report
parent
reply
2 points

You go, and read the citations.

permalink
report
parent
reply
4 points

Even with early GPT-4 it would also cite real citations that weren’t actually about the topic. So you may be doing a lot of work double checking as opposed to just looking into an answer yourself from the start.

permalink
report
parent
reply
2 points

I didn’t mean to cause any confusion, but what I said before was utter bullshit.

permalink
report
parent
reply
8 points

Isn’t this just a restatement of the halting problem?

permalink
report
reply

ChatGPT

!chatgpt@lemmy.world

Create post

Unofficial ChatGPT community to discuss anything ChatGPT

Community stats

  • 243

    Monthly active users

  • 296

    Posts

  • 2.3K

    Comments

Community moderators