You are viewing a single thread.
View all comments View context
2 points

TBF, compression is related to ML. Hence, the Hutter Prize. Thinking of LLMs as lossy compression algorithms is a decent analogy.

permalink
report
parent
reply
0 points

It is a partial analogy, it takes into consideration the outputs which are related to some specific training data and disconsiders the outputs which cannot be directly related to any specific training data.

For example, make up a new meme template and a new joke on the spot, it couldn’t have seen it before if you make sure your joke and template are new. If the AI can explain it then compression is a horrendous analogy.

Lossy compression explains outputs being similar but not identical when trying to recover the original data, it doesn’t explain brand new content that makes sense standalone. Imagine a lossy audio compression resulting in a brand new song midway through playback, or a lossy image compression resulting in a brand new coherent image being overlayed onto some pixels of the original image. That is not what happens, lossy audio compression results in noise, lossy image compression results in noise, not in coherent unheard songs and unseen images.

permalink
report
parent
reply

Funny

!funny@sh.itjust.works

Create post

General rules:

  • Be kind.
  • All posts must make an attempt to be funny.
  • Obey the general sh.itjust.works instance rules.
  • No politics or political figures. There are plenty of other politics communities to choose from.
  • Don’t post anything grotesque or potentially illegal. Examples include pornography, gore, animal cruelty, inappropriate jokes involving kids, etc.

Exceptions may be made at the discretion of the mods.

Community stats

  • 4.4K

    Monthly active users

  • 1.1K

    Posts

  • 19K

    Comments