164 points

Mhm I have mixed feelings about this. I know that this entire thing is fucked up but isn’t it better to have generated stuff than having actual stuff that involved actual children?

permalink
report
reply
112 points

A problem that I see getting brought up is that generated AI images makes it harder to notice photos of actual victims, making it harder to locate and save them

permalink
report
parent
reply
4 points

And doesn’t the AI learn from real images?

permalink
report
parent
reply
24 points

It does learn from real images, but it doesn’t need real images of what it’s generating to produce related content.
As in, a network trained with no exposure to children is unlikely to be able to easily produce quality depictions of children. Without training on nudity, it’s unlikely to produce good results there as well.
However, if it knows both concepts it can combine them readily enough, similar to how you know the concept of “bicycle” and that of “Neptune” and can readily enough imagine “Neptune riding an old fashioned bicycle around the sun while flaunting it’s tophat”.

Under the hood, this type of AI is effectively a very sophisticated “error correction” system. It changes pixels in the image to try to “fix it” to matching the prompt, usually starting from a smear of random colors (static noise).
That’s how it’s able to combine different concepts from a wide range of images to create things it’s never seen.

permalink
report
parent
reply
7 points

Basically if I want to create … (I’ll use a different example for obvious reasons, but I’m sure you could apply it to the topic)

… “an image of a miniature denium airjet with Taylor Swift’s face on the side of it”, the AI generators can despite no such thing existing in the training data. It may take multiple attempts and effort with the text prompt to get exactly what you’re looking for, but you could eventually get a convincing image.

AI takes loads of preexisting data on airplanes, T.Swift, and denium to combine it all into something new.

permalink
report
parent
reply
4 points

True, but by their very nature their generations tend to create anonymous identities, and the sheer amount of them would make it harder for investigators to detect pictures of real, human victims (which can also include indicators of crime location.

permalink
report
parent
reply
-4 points

Well that, and the idea of cathartic relief is increasingly being dispelled. Behaviour once thought to act as a pressure relief for harmful impulsive behaviour is more than likely just a pattern of escalation.

permalink
report
parent
reply
13 points

Source? From what I’ve heard, recent studies are showing the opposite.

permalink
report
parent
reply
72 points

The arrest is only a positive. Allowing pedophiles to create AI CP is not a victimless crime. As others point out it muddies the water for CP of real children, but it also potentially would allow pedophiles easier ways to network in the open (if the images are legal they can easily be platformed and advertised), and networking between abusers absolutely emboldens them and results in more abuse.

As a society we should never allow the normalization of sexualizing children.

permalink
report
parent
reply
39 points

Interesting. What do you think about drawn images? Is there a limit to how will the artist can be at drawing/painting? Stick figures vs life like paintings. Interesting line to consider.

permalink
report
parent
reply
17 points

If it was photoreal and difficult to distinguish from real photos? Yes, it’s exactly the same.

And even if it’s not photo real, communities that form around drawn child porn are toxic and dangerous as well. Sexualizing children is something I am 100% against.

permalink
report
parent
reply
-5 points

I’m in favor of specific legislation criminalizing drawn CSAM. It’s definitely less severe than photographic CSAM, and it’s definitely harmful.

permalink
report
parent
reply
28 points

networking between abusers absolutely emboldens them and results in more abuse.

Is this proven or a common sense claim you’re making?

permalink
report
parent
reply
-1 points
*

I wouldn’t be surprised if it’s a mixture of the two. It’s kind of like if you surround yourself with criminals regularly, you’re more likely to become one yourself. Not to say it’s a 100% given, just more probable.

permalink
report
parent
reply
-2 points

The far right in France normalized its discourses and they are now at the top of the votes.

Also in France, people talked about pedophilia at the TV in the 70s, 80s and at the beginning of the 90s. It was not just once in a while. It was frequent and open without any trouble. Writers would casually speak about sexual relationships with minors.

The normalization will blur the limits between AI and reality for the worse. It will also make it more popular.

The other point is also that people will always ends with the original. Again, politic is a good example. Conservatives try to mimic the far right to gain votes but at the end people vote for the far right…

And, someone has a daughter. A pedophile takes a picture of her without asking and ask an AI to produce CP based on her. I don’t want to see things like this.

permalink
report
parent
reply
18 points
*

Actually, that’s not quite as clear.

The conventional wisdom used to be, (normal) porn makes people more likely to commit sexual abuse (in general). Then scientists decided to look into that. Slowly, over time, they’ve become more and more convinced that (normal) porn availability in fact reduces sexual assault.

I don’t see an obvious reason why it should be different in case of CP, now that it can be generated.

permalink
report
parent
reply
2 points

I wonder if religiosity is correlated.

permalink
report
parent
reply
-1 points

It should be different because people can not have it. It is disgusting, makes them feel icky and thats just why it has to be bad. Conventional wisdom sometimes really is just convential idiocracy.

permalink
report
parent
reply
28 points

Did we memory hole the whole ‘known CSAM in training data’ thing that happened a while back? When you’re vacuuming up the internet you’re going to wind up with the nasty stuff, too. Even if it’s not a pixel by pixel match of the photo it was trained on, there’s a non-zero chance that what it’s generating is based off actual CSAM. Which is really just laundering CSAM.

permalink
report
parent
reply
31 points

IIRC it was something like a fraction of a fraction of 1% that was CSAM, with the researchers identifying the images through their hashes but they weren’t actually available in the dataset because they had already been removed from the internet.

Still, you could make AI CSAM even if you were 100% sure that none of the training images included it since that’s what these models are made for - being able to combine concepts without needing to have seen them before. If you hold the AI’s hand enough with prompt engineering, textual inversion and img2img you can get it to generate pretty much anything. That’s the power and danger of these things.

permalink
report
parent
reply
-5 points

What % do you think was used to generate the CSAM, though? Like, if 1% of the images were cups it’s probably drawing on some of that to generate images of cups.

And yes, you could technically do this with no CSAM training material, but we don’t know if that’s what the AI is doing because the image sources used to train it were mass scraped from the internet. They’re using massive amounts of data without filtering it and are unable to say with certainty whether or not there is CSAM in the training material.

permalink
report
parent
reply
10 points

I didn’t know that, my bad.

permalink
report
parent
reply
7 points

Fair but depressing, it seems like it barely registered in the news cycle.

permalink
report
parent
reply
26 points

Yeah, it’s very similar to the “is loli porn unethical” debate. No victim, it could supposedly help reduce actual CSAM consumption, etc… But it’s icky so many people still think it should be illegal.

There are two big differences between AI and loli though. The first is that AI would supposedly be trained with CSAM to be able to generate it. An artist can create loli porn without actually using CSAM references. The second difference is that AI is much much easier for the layman to create. It doesn’t take years of practice to be able to create passable porn. Anyone with a decent GPU can spin up a local instance, and be generating within a few hours.

In my mind, the former difference is much more impactful than the latter. AI becoming easier to access is likely inevitable, so combatting it now is likely only delaying the inevitable. But if that AI is trained on CSAM, it is inherently unethical to use.

Whether that makes the porn generated by it unethical by extension is still difficult to decide though, because if artists hate AI, then CSAM producers likely do too. Artists are worried AI will put them out of business, but then couldn’t the same be said about CSAM producers? If AI has the potential to run CSAM producers out of business, then it would be a net positive in the long term, even if the images being created in the short term are unethical.

permalink
report
parent
reply
23 points

Just a point of clarity, an AI model capable of generating csam doesn’t necessarily have to be trained on csam.

permalink
report
parent
reply
-1 points

That honestly brings up more questions than it answers.

permalink
report
parent
reply
3 points

I think one of the many problems with AI generated CSAM is that as AI becomes more advanced it will become increasingly difficult for authorities to tell the difference between what was AI generated and what isn’t.

Banning all of it means authorities don’t have to sift through images trying to decipher between the two. If one image is declared to be AI generated and it’s not…well… that doesn’t help the victims or create less victims. It could also make the horrible people who do abuse children far more comfortable putting that stuff out there because it can hide amongst all the AI generated stuff. Meaning authorities will have to go through far more images before finding ones with real victims in it. All of it being illegal prevents those sorts of problems.

permalink
report
parent
reply
2 points
*

And that’s a good point! Luckily it’s still (usually) fairly easy to identify AI generated images. But as they get more advanced, that will likely become harder and harder to do.

Maybe some sort of required digital signatures for AI art would help; Something like a public encryption key in the metadata, that can’t be falsified after the fact. Anything without that known and trusted AI signature would by default be treated as the real deal.

But this would likely require large scale rewrites of existing image formats, if they could even support it at all. It’s the type of thing that would require people way smarter than myself. But even that feels like a bodged solution to a problem that only exists because people suck. And if it required registration with a certificate authority (like an HTTPS certificate does) then it would be a hurdle for local AI instances to jump through. Because they would need to get a trusted certificate before they could sign their images.

permalink
report
parent
reply
1 point
*

But it’s icky so many people still think it should be illegal.

Imo, not the best framework for creating laws. Essentially, it’s an appeal to emotion.

permalink
report
parent
reply
-13 points

so many people still think it should be illegal

It is illegal. https://www.thefederalcriminalattorneys.com/possession-of-lolicon

permalink
report
parent
reply
15 points
*

I wasn’t arguing about current laws. I was simply arguing about public perception, and whether the average person believes it should be illegal. There’s a difference between legality and ethicality. Something unethical can be legal, and something illegal can be ethical.

Weed is illegal, but public perception says it shouldn’t be.

permalink
report
parent
reply
4 points

Why are you assuming everyone lives in the US? Your article even admits that it is legal elsewhere (Japan).

This is a better one: https://en.wikipedia.org/wiki/Legal_status_of_fictional_pornography_depicting_minors

permalink
report
parent
reply
11 points

I have trouble with this because it’s like 90% grey area. Is it a pic of a real child but inpainted to be nude? Was it a real pic but the face was altered as well? Was it completely generated but from a model trained on CSAM? Is the perceived age of the subject near to adulthood? What if the styling makes it only near realistic (like very high quality CG)?

I agree with what the FBI did here mainly because there could be real pictures among the fake ones. However, I feel like the first successful prosecution of this kind of stuff will be a purely moral judgement of whether or not the material “feels” wrong, and that’s no way to handle criminal misdeeds.

permalink
report
parent
reply
15 points

If not trained on CSAM or in painted but fully generated, I can’t really think of any other real legal arguments against it except for: “this could be real”. Which has real merit, but in my eyes not enough to prosecute as if it were real. Real CSAM has very different victims and abuse so it needs different sentencing.

permalink
report
parent
reply
2 points

Everything is 99% grey area. If someone tells you something is completely black and white you should be suspicious of their motives.

permalink
report
parent
reply
9 points

You know whats better? Having none of this shit

permalink
report
parent
reply
16 points

Did you just fix menal health?

permalink
report
parent
reply
8 points

Yeah as I also said.

permalink
report
parent
reply
2 points
*

Nirvana fallacy

Yeah would be nice. Unfortunelately it isn’t so and it’s never going to. Chasing after people generating distasteful AI pictures is not making the world a better place.

permalink
report
parent
reply
-3 points

Better for whom and why?

permalink
report
parent
reply
9 points

Apparently he sent some to an actual minor.

permalink
report
parent
reply
3 points

It reminds me of the story of the young man who realized he had an attraction to underage children and didn’t want to act on it, yet there were no agencies or organizations to help him, and that it was only after crimes were committed that anyone could get help.

I see this fake cp as only a positive for those people. That it might make it difficult to find real offenders is a terrible reason against.

permalink
report
parent
reply
2 points

Better only means less worse in this case, I guess

permalink
report
parent
reply
0 points

I think the point is that child attraction itself is a mental illness and people indulging it even without actual child contact need to be put into serious psychiatric evaluation and treatment.

permalink
report
parent
reply
-1 points

Yes, but the perp showed the images to a minor.

permalink
report
parent
reply
-2 points
*
Removed by mod
permalink
report
parent
reply
4 points

No?

permalink
report
parent
reply
1 point

Is everything completely black and white for you?

The system isn’t perfect, especially where we prioritize punishing people over rehabilitation. Would you rather punish everyone equally, emphasizing that if people are going to risk the legal implications (which, based on legal systems the world over, people are going to do) they might as well just go for the real thing anyways?

You don’t have to accept it as morally acceptable, but you don’t have to treat them as completely equivalent either.

There’s gradations of questionable activity. Especially when there’s no real victims involved. Treating everything exactly the same is, frankly speaking, insane. Its like having one punishment for all illegal behavior. Murder someone? Death penalty. Rob them? Straight to the electric chair. Jaywalking? Better believe you’re getting the needle.

permalink
report
parent
reply
-1 points
*

Ironically, You ask if everything is completely black and white for someone without accepting that there’s nuance to the very issue you’re calling out. And assuming that “everything”- a very black and white term, is not very nuanced, is it?

No, not EVERYTHING, but some things. And this is one of those things. Both forms should be illegal. Period. No nuance, no argument, NO grey area.

This does not mean that nuance doesn’t exist. It just means that some believe that it SHOULDN’T exist within the paradigm of child porn.

permalink
report
parent
reply
-1 points
*
Removed by mod
permalink
report
parent
reply
-4 points

It’s better to have neither.

permalink
report
parent
reply
7 points

This mentality smells of “just say no” for drugs or “just don’t have sex” for abortions. This is not the ideal world and we have to find actual plans/solutions to deal with the situation. We can’t just cover our ears and hope people will stop

permalink
report
parent
reply
-5 points

It feeds and evolves a disorder which in turn increases risks of real life abuse.

But if AI generated content is to be considered illegal, so should all fictional content.

permalink
report
parent
reply
24 points
*

Or, more likely, it feeds and satisfies a disorder which in turn decreases risk of real life abuse.

Making it illegal so far helped nothing, just like with drugs

permalink
report
parent
reply
-5 points

That’s not how these addictive disorders works… they’re never satisfied and always need more.

permalink
report
parent
reply
5 points

Two things:

  1. Do we know if fuels the urge to get real children? Or do we just assume that through repetition like the myth of “gateway drugs”?
  2. Since no child was involved and harmed in the making of these images… On what grounds could it be forbidden to generate them?
permalink
report
parent
reply
4 points
*

Alternative perspective is to think that does watching normal porn make heterosexual men more likely to rape women? If not then why would it be different in this case?

The vast majority of pedophiles never offend. Most people in jail for child abuse are just plain old rapists with no special interest towards minors, they’re just an easy target. Pedophilia just describes what they’re attracted to. It’s not a synonym to child rapist. It usually needs to coinside with psychopathy to create the monster that most people think about when hearing that word.

permalink
report
parent
reply
0 points

I would love to see research data pointing either way re #1, although it would be incredibly difficult to do so ethically, verging on impossible. For #2, people have extracted originals or near-originals of inputs to the algorithms. AI generated stuff - plagiarism machine generated stuff, runs the risk of effectively revictimizing people who were already abused to get said inputs.

It’s an ugly situation all around, and unfortunately I don’t know that much can be done about it beyond not demonizing people who have such drives, who have not offended, so that seeking therapy for the condition doesn’t screw them over. Ensuring that people are damned if they do and damned if they don’t seems to pretty reliably produce worse outcomes.

permalink
report
parent
reply
151 points
*

The headline/title needs to be extended to include the rest of the sentence

“and then sent them to a minor”

Yes, this sicko needs to be punished. Any attempt to make him the victim of " the big bad government" is manipulative at best.

Edit: made the quote bigger for better visibility.

permalink
report
reply
50 points

That’s a very important distinction. While the first part is, to put it lightly, bad, I don’t really care what people do on their own. Getting real people involved, and minor at that? Big no-no.

permalink
report
parent
reply
26 points

All LLM headlines are like this to fuel the ongoing hysteria about the tech. It’s really annoying.

permalink
report
parent
reply
8 points
*

Sure is. I report the ones I come across as clickbait or missleading title, explaining the parts left out…such as this one where those 7 words change the story completely.

Whoever made that headline should feel ashamed for victimizing a grommer.

permalink
report
parent
reply
7 points

I’d be torn on the idea of AI generating CP, if it were only that. On one hand if it helps them calm the urges while no one is getting hurt, all the better. But on the other hand it might cause them not to seek help, but problem is already stigmatized severely enough that they are most likely not seeking help anyway.

But sending that stuff to a minor. Big problem.

permalink
report
parent
reply
-1 points

Cartoon CSAM is illegal in the United States. Pretty sure the judges will throw his images under the same ruling.

https://en.wikipedia.org/wiki/PROTECT_Act_of_2003

https://www.thefederalcriminalattorneys.com/possession-of-lolicon

permalink
report
parent
reply
9 points

It won’t. They’ll get them for the actual crime not the thought crime that’s been nerfed to oblivion.

permalink
report
parent
reply
3 points

Based on the blacklists that one has to fire up before browsing just about any large anime/erotica site, I am guessing that these “laws” are not enforced, because they are flimsy laws to begin with. Reading the stipulations for what constitutes a crime is just a hotbed for getting an entire case tossed out of court. I doubt any prosecutors would lean hard on possession of art unless it was being used in another crime.

permalink
report
parent
reply
97 points

Bad title.

They caught him not simply for creating pics, but also for trading such pics etc.

permalink
report
reply
5 points

You can get away with a lot of heinous crimes by simply not telling people and sharing the results.

permalink
report
parent
reply
1 point

You consider it a heinous crime to draw a picture and keep it to yourself?

permalink
report
parent
reply
-5 points

Read the article. He was arrested for sending the pictures to at least one minor.

permalink
report
parent
reply
-3 points
permalink
report
parent
reply
-12 points

That’s sickening to know there are bastards out there who will get away with it since they are only creating it.

permalink
report
parent
reply
12 points

I’m not sure. Let us assume that you generate it on your own PC at home (not using a public service) and don’t brag about it and never give it to anybody - what harm is done?

permalink
report
parent
reply
4 points
*

Society is not ok with the idea of someone cranking to CSAM, then just walking around town. It gives people wolf-in-sheep-clothing vibes.

So the notion of there being “ok” CSAM-style ai content is a non starter for a huge fraction of people because it still suggests appeasing a predator.

I’m definitely one of those people that simply can’t accept any version of it.

permalink
report
parent
reply
-13 points

Even if the AI didn’t train itself on actual CSAM that is something that feels inherently wrong. Your mind is not right to think that’s acceptable IMO.

permalink
report
parent
reply
61 points

It’s worth mentioning that in this instance the guy did send porn to a minor. This isn’t exactly a cut and dry, “guy used stable diffusion wrong” case. He was distributing it and grooming a kid.

The major concern to me, is that there isn’t really any guidance from the FBI on what you can and can’t do, which may lead to some big issues.

For example, websites like novelai make a business out of providing pornographic, anime-style image generation. The models they use deliberately tuned to provide abstract, “artistic” styles, but they can generate semi realistic images.

Now, let’s say a criminal group uses novelai to produce CSAM of real people via the inpainting tools. Let’s say the FBI cast a wide net and begins surveillance of novelai’s userbase.

Is every person who goes on there and types, “Loli” or “Anya from spy x family, realistic, NSFW” (that’s an underaged character) going to get a letter in the mail from the FBI? I feel like it’s within the realm of possibility. What about “teen girls gone wild, NSFW?” Or “young man, no facial body hair, naked, NSFW?”

This is NOT a good scenario, imo. The systems used to produce harmful images being the same systems used to produce benign or borderline images. It’s a dangerous mix, and throws the whole enterprise into question.

permalink
report
reply
10 points

The major concern to me, is that there isn’t really any guidance from the FBI on what you can and can’t do, which may lead to some big issues.

https://www.ic3.gov/Media/Y2024/PSA240329 https://www.justice.gov/criminal/criminal-ceos/citizens-guide-us-federal-law-child-pornography

They’ve actually issued warnings and guidance, and the law itself is pretty concise regarding what’s allowed.

(8) “child pornography” means any visual depiction, including any photograph, film, video, picture, or computer or computer-generated image or picture, whether made or produced by electronic, mechanical, or other means, of sexually explicit conduct, where-

(A) the production of such visual depiction involves the use of a minor engaging in sexually explicit conduct;

(B) such visual depiction is a digital image, computer image, or computer-generated image that is, or is indistinguishable from, that of a minor engaging in sexually explicit conduct; or

© such visual depiction has been created, adapted, or modified to appear that an identifiable minor is engaging in sexually explicit conduct.

(11) the term “indistinguishable” used with respect to a depiction, means virtually indistinguishable, in that the depiction is such that an ordinary person viewing the depiction would conclude that the depiction is of an actual minor engaged in sexually explicit conduct. This definition does not apply to depictions that are drawings, cartoons, sculptures, or paintings depicting minors or adults.

https://uscode.house.gov/view.xhtml?hl=false&edition=prelim&req=granuleid%3AUSC-prelim-title18-section2256&f=treesort&num=0

If you’re going to be doing grey area things you should do more than the five minutes of searching I did to find those honestly.

It was basically born out of a supreme Court case in the early 2000s regarding an earlier version of the law that went much further and banned anything that “appeared to be” or “was presented as” sexual content involving minors, regardless of context, and could have plausibly been used against young looking adult models, artistically significant paintings, or things like Romeo and Juliet, which are neither explicit nor vulgar but could be presented as involving child sexual activity. (Juliet’s 14 and it’s clearly labeled as a love story).
After the relevant provisions were struck down, a new law was passed that factored in the justices rationale and commentary about what would be acceptable and gave us our current system of “it has to have some redeeming value, or not involve actual children and plausibly not look like it involves actual children”.

permalink
report
parent
reply
-6 points

The major concern to me, is that there isn’t really any guidance from the FBI on what you can and can’t do, which may lead to some big issues.

The Protect Act of 2003 means that any artistic depiction of CSAM is illegal. The guidance is pretty clear, FBI is gonna raid your house…eventually. We still haven’t properly funded the anti-CSAM departments.

permalink
report
parent
reply
-20 points

Is every person who goes on there and types, “Loli” or “Anya from spy x family, realistic, NSFW” (that’s an underaged character) going to get a letter in the mail from the FBI?

I’ll throw that baby out with the bathwater to be honest.

permalink
report
parent
reply
33 points
*

Simulated crimes aren’t crimes. Would you arrest every couple that finds health ways to simulate rape fetishes? Would you arrest every person who watches Fast and The Furious or The Godfather?

If no one is being hurt, if no real CSAM is being fed into the model, if no pornographic images are being sent to minors, it shouldn’t be a crime. Just because it makes you uncomfortable, don’t make it immoral.

permalink
report
parent
reply
26 points

Or, ya know, everyone who ever wanted to decapitate those stupid fucking Skyrim children. Crime requires damaged parties, and with this (idealized case, not the specific one in the article) there is none.

permalink
report
parent
reply
21 points

Simulated crimes aren’t crimes.

If they were, any one who’s played games is fucked. I’m confident everyone who has played went on a total ramapage murdering the townfolk, pillaging their houses and blowing everything up…in Minecraft.

permalink
report
parent
reply
3 points

They would though. We know they would because conservatives already did the whole laws about how you can have sex in private thing.

permalink
report
parent
reply
-8 points

Simulated crimes aren’t crimes.

Artistic CSAM is definitely a crime in the United States. PROTECT act of 2003.

permalink
report
parent
reply
-9 points
*
Deleted by creator
permalink
report
parent
reply
-1 points
Deleted by creator
permalink
report
parent
reply
44 points
*

America has some of the most militant anti pedophilic culture in the world but they far and away have the highest rates of child sexual assault.

I think AI is going to revel is how deeply hypocritical Americans are on this issue. You have gigantic institutions like churches committing industrial scale victimization yet you won’t find a 1/10th of the righteous indignation against other organized religions where there is just as much evidence it is happening as you will regarding one person producing images that don’t actually hurt anyone.

It’s pretty clear by how staggering a rate of child abuse that occurs in the states that Americans are just using child victims as weaponized politicalization (it’s next to impossible to convincingly fight off pedo accusations if you’re being mobbed) and aren’t actually interested in fighting pedophilia.

permalink
report
reply
26 points

Most states will let grown men marry children as young as 14. There is a special carve out for Christian pedophiles.

permalink
report
parent
reply
2 points

Fortunately most instances are in the category of a 17 year old to an 18 year old, and require parental consent and some manner of judicial approval, but the rates of “not that” are still much higher than one would want.
~300k in a 20 year window total, 74% of the older partner being 20 or younger, and 95% of the younger partner being 16 or 17, with only 14% accounting for both partners being under 18.

There’s still no reason for it in any case, and I’m glad to live in one of the states that said "nah, never needed .

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 10K

    Posts

  • 457K

    Comments