You are viewing a single thread.
View all comments View context
-2 points

That’s literally how they work

permalink
report
parent
reply
4 points

Man the models can’t store verbatim its training data, the amount of data is turned into a model that is hundreds or thousands of times smaller than the original source data. If it was capable of simply recovering everything that it was trained on this would be some magical compression algorithm and that by itself would be extremely impressive.

permalink
report
parent
reply
-4 points

Congratulations on discovering compression

permalink
report
parent
reply
2 points
*

Oh ok, you want to claim this is compressing the entirety of the internet in a model that isn’t even 1 terabyte of data and be unimpressed that is something.

But it isn’t compression. It is a mathematical fact that neural networks are universal function approximators, this is undisputed, and analytic functions are continuous so to be an analytical function approximator it must be able to fill in the gaps between discrete data points by itself, which necessarily means spiting out data outside of the input distribution, data it has not seen.

permalink
report
parent
reply
2 points

They do not store anything verbatim; They instead store the directions in which various words and related concepts relate to one another in some gigantic multidimensional space.

I highly suggest you go learn what they actually do before you continue talking out of your ass about them

permalink
report
parent
reply
0 points

If you trained a GPT on a single phrase, all you’d get out of it would be the single phrase.

The mechanism of storage doesn’t need to be just the verbatim source material, which is not even close to what I said.

permalink
report
parent
reply
1 point
*

You said it matches text to its training data, which it does not do.

Your single-phrase statement only works for very short, non-repetitive phrases. As soon as your phrase repeats a token more than a few times, the statistics for the tokens change and could result in nonsensical output that repeats through subsections of the training data.

And even then for that single non-repetitive phrases, the reason you would get that single phrase back is not because it would be “matching on” the phrase. It is because the token weights would effectively encode that the statistical likelihood of the “next token” in the generated output is 100% for a given token when the evaluated token precedes it in the training phrase. Or in other words: Your training data being a single phrase maniplates the statistics so that the most likely output is that single phrase.

However, that is a far cry from simple “matching” against the training data. Which is what you said it does.

permalink
report
parent
reply

Funny

!funny@sh.itjust.works

Create post

General rules:

  • Be kind.
  • All posts must make an attempt to be funny.
  • Obey the general sh.itjust.works instance rules.
  • No politics or political figures. There are plenty of other politics communities to choose from.
  • Don’t post anything grotesque or potentially illegal. Examples include pornography, gore, animal cruelty, inappropriate jokes involving kids, etc.

Exceptions may be made at the discretion of the mods.

Community stats

  • 4.4K

    Monthly active users

  • 1.1K

    Posts

  • 19K

    Comments