You are viewing a single thread.
View all comments
59 points

Whenever I stumble on reddit I make sure to post disinformation or some kind of dumb shit to throw a wrench into the LLM training data they sell to google.

permalink
report
reply
21 points

I have literally over a hundred permabanned accounts on that site.

permalink
report
parent
reply
8 points

That’s impressive, I think I’ve got half a dozen.

permalink
report
parent
reply
4 points

I timed it once, took me just over a minute to make a new account.

permalink
report
parent
reply
3 points

I just got another one today for “harassment” of zionism in r/worldnews. Reddit cannot hold a free discussion and they know it. They can’t even let you speak to expose their bullshit, and permaban you when you do.

I can you show you the comment which got me banned. I was literally asking questions which they know the answer for but censor intentionally because they are bought and controlled by awful groups directly linked to the IDF themselves. They have a division who train and employe teens as stupid Hasbara trolls who don’t know history and unable to hold a discussion.

permalink
report
parent
reply
5 points

My first one was in response to some rich cunt who went on air to say more or less that poverty was a good thing because then people had something to strive to avoid, so I said something along the lines of “This guy should be shot, I’m not even exaggerating.”

permalink
report
parent
reply
15 points

Make sure to have some LLM generate the comment for you, as LLMs learning synthetic data may fuck them up over time: AI models fed AI-generated data quickly spew nonsense

permalink
report
parent
reply
3 points
*

I hate to ruin this for you, but if you post nonsense, it will get downvoted by humans and excluded from any data set (or included as examples of what to avoid). If it’s not nonsensical enough to be downvoted, it still won’t do well vote wise, and will not realistically poison any data. And if it’s upvoted… it just might be good data. That is why Reddit’s data is valuable to Google. It basically has a built in system for identifying ‘bad’ data.

permalink
report
parent
reply
1 point

No, you’re missing the point. You make up some credible misinformation to poison AI training with, but you don’t stop there: you get an LLM to rewrite it for you. Retry until you get a text that sounds credible, doesn’t particularly look written by AI, and people will upvote, and post that.

With this, even if the text looks good, you’re not only poisoning future models with the misinformation you started with; by feeding them a text generated by an LLM (even if you can’t tell the difference at first glance) you’re introducing feedback into the model that will also poison it, not with misinformation, but by reinforcing its biases and errors and reducing the variety of its training data.

permalink
report
parent
reply
1 point

this is an ancient and noble practice known as shitposting, no need to call it something else :)

permalink
report
parent
reply

Greentext

!greentext@sh.itjust.works

Create post

This is a place to share greentexts and witness the confounding life of Anon. If you’re new to the Greentext community, think of it as a sort of zoo with Anon as the main attraction.

Be warned:

  • Anon is often crazy.
  • Anon is often depressed.
  • Anon frequently shares thoughts that are immature, offensive, or incomprehensible.

If you find yourself getting angry (or god forbid, agreeing) with something Anon has said, you might be doing it wrong.

Community stats

  • 6.9K

    Monthly active users

  • 997

    Posts

  • 40K

    Comments