Small rant : Basically, the title. Instead of answering every question, if it instead said it doesn’t know the answer, it would have been trustworthy.

15 points

This wasn’t an intentional feature; they’re actually trying to train it with fine-tuning to add this as an ability. It’s one area that highlights the difference between it imitating the text it’s been seeing, instead of actually understanding what it’s saying – since most of its training data is of the form “(ask a question) (response to question)” overwhelmingly more often than “(ask a question) (say you don’t know, the end)”, it is trying to be a good imitator and do the same, and come up with some plausible nonsense even if it doesn’t know the answer.

permalink
report
reply
7 points

And sometimes that’s exactly what I want, too. I use LLMs like ChatGPT when brainstorming and fleshing out fictional scenarios for tabletop roleplaying games, for example, and in those situations coming up with plausible nonsense is specifically the job at hand. I wouldn’t want to go “ChatGPT, I need a description of the interior of a wizard’s tower is like” and get the response “I don’t know what the interior of a wizard’s tower is like.”

permalink
report
parent
reply
7 points

At one point I messed around with a lore generator that would chop up sections of “The Dungeon Alphabet” and “Fire on the Velvet Horizon” along with some other stuff, and feed random sections of them into the LLM for inspiration and then ask it to lay out a little map, and it pretty reliably came up with all kind of badass stuff.

permalink
report
parent
reply
3 points

Part of the problem is fine tuning is very shallow, and that a contributing issue for claiming to be right when it isn’t is the pretraining on a bunch of training data of people online claiming to be right when they aren’t.

permalink
report
parent
reply
1 point

Yeah. It is fairly weird to me that it’s such a common thing to do to take the raw output of the LLM and send that to the user, and to try use fine-tuning to get that raw output to look some way that you want.

To me it is obvious that something like having the LLM emit a little JSON block which includes some field which covers “how sure are you that this is actually true” or something, is more flexible and simpler and cheaper and works better.

But what do I know

permalink
report
parent
reply
1 point

The problem is that they are prone to making up why they are correct too.

There’s various techniques to try and identify and correct hallucinations, but they all increase the cost and none are a silver bullet.

But the rate at which it occurs decreased with the jump in pretrained models, and will likely decrease further with the next jump too.

permalink
report
parent
reply
1 point
*

Good look getting it to reply consistently with a json object

Edit: maybe i’m shit at prompting but for me it’s almost impossible to even get it to just shut up and consistently reply yes or no to my questions

permalink
report
parent
reply
2 points

Doesnt the bot already imply that it could be wrong?

permalink
report
reply
4 points

I work with plenty of people who don’t even do that. They just keep making stuff up like they do… But they’re confident in their incorrect answers, so people listen to them.

permalink
report
reply
13 points

Even a response that it doesn’t know an answer would be untrustworthy

permalink
report
reply
23 points

That would require ChatGPT to know that it’s talking bullshit. It’s not a knowledge database, it’s a digital parrot.

permalink
report
reply

ChatGPT

!chatgpt@lemmy.world

Create post

Unofficial ChatGPT community to discuss anything ChatGPT

Community stats

  • 678

    Monthly active users

  • 276

    Posts

  • 2.2K

    Comments

Community moderators