194 points

Collapse faster, please. Sick of ai bullshit clogging up my searches.

permalink
report
reply
90 points

The Internet is fucked now, the only valuable untainted training data is the Internet as it existed prior to this AI bullshit coming online. Confirmed human content is going to be super valuable, so expect our privacy to be fucked as well…

permalink
report
parent
reply
6 points

Did anyone take a copy?

permalink
report
parent
reply
8 points

The Internet Archive has some.

permalink
report
parent
reply
3 points

Even that is going to turn into a shit show … It will become a copy of a copy of a copy of a backup of a backup of a copy and all of it will just get rendered down to some common basics based on whatever the hell was marketed and promoted by bots

permalink
report
parent
reply
42 points
*

The collapse won’t stop ai output from spamming the internet though. It will just make it worse and more likely to be incorrect

permalink
report
parent
reply
4 points

We can just stop checking the internet for things

permalink
report
parent
reply
1 point

Really though, is does anyone know if/how Wikipedia is protecting itself from this?

permalink
report
parent
reply
1 point

It’s not going to. It’s just going to get more widespread and harder to detect. The incentives favor developing better and better AI. Luckily one of the solutions to this issue is - wait for it - AI. With a good enough AI, especially a generally intelligent one you don’t need search engines anymore. You just ask and it gives you the answer. If you think AI couldn’t do this reliably then that is not the AI I’m talking about.

permalink
report
parent
reply
76 points

My team has been calling models that use ai generated data “Habsberg models”

permalink
report
reply
20 points

I feel there is a good joke here, but I miss the knowledge to understand it. Care to enlighten me?

permalink
report
parent
reply
60 points

The Habsburg royal family line was famously inbred with a distinctive chin.

permalink
report
parent
reply
4 points

For anyone curious on how bad it got look up the coroners report on Charles of Spain. Its fucken grizzly.

permalink
report
parent
reply
12 points

Lmao that’s a perfect name for it

permalink
report
parent
reply
5 points

I thought it was called centipeding

permalink
report
parent
reply
67 points

Maybe we need to label AI-generated content to, you know, avoid confusion.

permalink
report
reply
111 points

Sorry, best we can do is a race to the bottom fueled by greed and incompetence.

permalink
report
parent
reply
44 points

That will be a refeshing change.

permalink
report
parent
reply
4 points

That’s what has been happening, and is likely what will continue to happen. Not much change there really…

permalink
report
parent
reply
6 points

I’m sure we can compromise on a mandatory database of registered AI-generated content that only the corporations can read from but everyone using AI-generated content is required by law to write to, with hefty fines (but only for regular people).

permalink
report
parent
reply
43 points
*

Oh goody. I’ve been wanting to use this since my slashdot days… today is my first chance!

Your post advocates a

[x] technical
[ ] legislative
[ ] market-based
[ ] vigilante

approach to fighting (ML-generated) spam. Your idea will not work. Here is why
it won't work. [One or more of the following may apply to your particular idea,
and it may have other flaws which used to vary from state to state before a bad
federal law was passed.]

[ ] Spammers can easily use it to harvest email addresses
[ ] Mailing lists and other legitimate email uses would be affected
[ ] No one will be able to find the guy or collect the money
[ ] It is defenseless against brute force attacks
[ ] It will stop spam for two weeks and then we'll be stuck with it
[ ] Users of email will not put up with it
[x] Microsoft will not put up with it
[ ] The police will not put up with it
[x] Requires too much cooperation from spammers
[x] Requires immediate total cooperation from everybody at once
[ ] Many email users cannot afford to lose business or alienate potential employers
[ ] Spammers don't care about invalid addresses in their lists
[ ] Anyone could anonymously destroy anyone else's career or business

Specifically, your plan fails to account for

[ ] Laws expressly prohibiting it
[x] Lack of centrally controlling authority for email^W ML algorithms
[ ] Open relays in foreign countries
[ ] Ease of searching tiny alphanumeric address space of all email addresses
[x] Asshats
[ ] Jurisdictional problems
[ ] Unpopularity of weird new taxes
[ ] Public reluctance to accept weird new forms of money
[ ] Huge existing software investment in SMTP
[ ] Susceptibility of protocols other than SMTP to attack
[ ] Willingness of users to install OS patches received by email
[ ] Armies of worm riddled broadband-connected Windows boxes
[x] Eternal arms race involved in all filtering approaches
[x] Extreme profitability of spam
[ ] Joe jobs and/or identity theft
[ ] Technically illiterate politicians
[ ] Extreme stupidity on the part of people who do business with spammers
[x] Dishonesty on the part of spammers themselves
[ ] Bandwidth costs that are unaffected by client filtering
[x] Outlook

and the following philosophical objections may also apply:

[x] Ideas similar to yours are easy to come up with, yet none have ever
been shown practical
[ ] Any scheme based on opt-out is unacceptable
[ ] SMTP headers should not be the subject of legislation
[ ] Blacklists suck
[ ] Whitelists suck
[ ] We should be able to talk about Viagra without being censored
[ ] Countermeasures should not involve wire fraud or credit card fraud
[ ] Countermeasures should not involve sabotage of public networks
[ ] Countermeasures must work if phased in gradually
[ ] Sending email should be free
[x] Why should we have to trust you and your servers?
[ ] Incompatiblity with open source or open source licenses
[x] Feel-good measures do nothing to solve the problem
[ ] Temporary/one-time email addresses are cumbersome
[ ] I don't want the government reading my email
[ ] Killing them that way is not slow and painful enough

Furthermore, this is what I think about you:

[x] Sorry dude, but I don't think it would work.
[ ] This is a stupid idea, and you're a stupid person for suggesting it.
[ ] Nice try, assh0le! I'm going to find out where you live and burn your
house down!
permalink
report
parent
reply
10 points
*

I traced this baby back to January 19th, 2004: https://craphound.com/spamsolutions.txt

permalink
report
parent
reply
2 points

Thanks, I was wondering how old it was when they said “Slashdot days.”

permalink
report
parent
reply
2 points

Oh, do me next, do me. Open source adversarial models trained to detect and actively label things which it detects as belonging to AI. Probably would end up looking like a browser extension or something. Ublock, but for AI, basically.

permalink
report
parent
reply
8 points

Sounds great, how do we enforce it?

permalink
report
parent
reply
14 points

If the AIs want to avoid digital incest they’ll enforce it for themselves.

permalink
report
parent
reply
1 point

The AIs dont want anything themselves and those who make the decisions about them want the most profit, what costs more, verifying training data or AI incest?

permalink
report
parent
reply
6 points

Sounds like something an advanced language learning model would say…

permalink
report
parent
reply
4 points

It’s important to understand that a language modelling AI can only produce responses based on its inputs.

permalink
report
parent
reply
4 points

Ah, you’re suggesting using RFC 3514. Good thinking.

permalink
report
parent
reply
1 point

Thank you for bringing that standard to my attention.

permalink
report
parent
reply
0 points

Far too late for that now.

permalink
report
parent
reply
40 points
*

Most people here don’t understand what this is saying.

We’ve had “pure” human generated data, verifiably so since LLMs and ImageGen didn’t exist. Any bot generated data was easily filterable due to lack of sophistication.

ChatGPT and SD3 enter the chat, generate nearly indistinguishable data from humans, but with a few errors here and there. These errors while few, are spectacular and make no sense to the training data.

2 years later, the internet is saturated with generated content. The old datasets are like gold now, since none of the new data is verifiably human.

This matters when you’ve played with local machine learning and understand how these machines “think”. If you feed an AI generated set to an AI as training data, it learns the mistakes as well as the data. Every generation it’s like mutations form until eventually it just produces garbage.

Training models on generated sets slowly by surely fail without a human touch. Scale this concept to the net fractionally. When 50% of your dataset is machine generated, 50% of your new model trained on it will begin to deteriorate. Do this long enough and that 50% becomes 60 to 70 and beyond.

Human creativity and thought have yet to be replicated. These models have no human ability to be discerning or sleep to recover errors. They simply learn imperfectly and generate new less perfect data in a digestible form.

permalink
report
reply
36 points

Looks like we need Low-Background-Content

permalink
report
reply
3 points

Can you explain further more ? Sorry I may understand but not sure

permalink
report
parent
reply
25 points

Low Background Radiation Steel was/is valuable, because it’s made of steel from before nuclear testing. As the bombs contaminated the produced steel.

In the same sense, anything before the creation of LLMs would be considered “low background radiation” content, as that’s the only content to be sure to be made without LLMs in the loop

permalink
report
parent
reply
5 points

I like the concept and analogy, thanks for explanations 🙏

permalink
report
parent
reply
-5 points

I can just take a picture with my phone whenever I want. This is a bad analogy

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 12K

    Posts

  • 538K

    Comments