Dear Lemmy.world Community,

Recently posts were made to the AskLemmy community that go against not just our own policies but the basic ethics and morals of humanity as a whole. We acknowledge the gravity of the situation and the impact it may have had on our users. We want to assure you that we take this matter seriously and are committed to making significant improvements to prevent such incidents in the future. Considering I’m reluctant to say exactly what these horrific and repugnant images were, I’m sure you can probably guess what we’ve had to deal with and what some of our users unfortunately had to see. I’ll add the thing we’re talking about in spoilers to the end of the post to spare the hearts and minds of those who don’t know.

Our foremost priority is the safety and well-being of our community members. We understand the need for a swift and effective response to inappropriate content, and we recognize that our current systems, protocols and policies were not adequate. We are immediately taking immediate steps to strengthen our moderation and administrative teams, implementing additional tools, and building enhanced pathways to ensure a more robust and proactive approach to content moderation. Not to mention ensuring ways that these reports are seen more quickly and succinctly by mod and admin teams.

The first step will be limiting the image hosting sites that Lemmy.world will allow. We understand that this can cause frustration for some of our users but we also hope that you can understand the gravity of the situation and why we find it necessary. Not just to protect all of our users from seeing this but also to protect ourselves as a site. That being said we would like input in what image sites we will be whitelisting. While we run a filter over all images uploaded to Lemmy.world itself, this same filter doesn’t apply to other sites which leads to the necessity of us having to whitelist sites.

This is a community made by all of us, not just by the admins. Which leads to the second step. We will be looking for more moderators and community members that live in more diverse time zones. We recognize that at the moment it’s relatively heavily based between Europe and North America and want to strengthen other time zones to limit any delays as much as humanly possible in the future.

We understand that trust is essential, especially when dealing with something as awful as this, and we appreciate your patience as we work diligently to rectify this situation. Our goal is to create an environment where all users feel secure and respected and more importantly safe. Your feedback is crucial to us, and we encourage you to continue sharing your thoughts and concerns.

Every moment is an opportunity to learn and build, even the darkest ones.

Thank you for your understanding.


Sincerely,

The Lemmy.world Administration

Legal / ToS

spoiler

CSAM


51 points

What is the complete correct response for users to carry out if they spot CP?

Is just the report button on the post good enough? Is there some kind of higher level report for bigger-than-just-one-instance shit that threatens Lemmy as a whole? Should we call the FBI or some shit?

I haven’t come across any here, but if I do, I’d like to be able to aid in swift action against not only the post/account in question, but against the actual person running it.

permalink
report
reply
26 points

In theory, reporting to the community moderators should be enough for users. It would then be the responsibility of thost moderators to report it to the instance admin, and then the admins responsibility to report it to the instance’s local law enforcement. They will then handle it appropriately.

However, sometimes community moderators are corrupt and will ignore reports and even ban users for reporting instance rule breaking content. In those cases, the user must report directly to the instance admin. As you can imagine, instance admins also can be corrupt and therefore the user must report to law enforcement.

But typically the first scenario is sufficient.

permalink
report
parent
reply
19 points

FYI: admins can see all reports. We currently have a tool running that scans for posts that are reported a lot, which will then notify people who can do something about it.

permalink
report
parent
reply
4 points

That seems too slow. Sibling comment by admin is reassuring here.

permalink
report
parent
reply
2 points
*
Deleted by creator
permalink
report
parent
reply
16 points

You can always report casm to the FBI. You don’t need a mod’s guidance to do so.

Otherwise report and block.

permalink
report
parent
reply
9 points
*

Report and block should be a correct respond.

Edit: actually also report the user to the instance admin where the user from.

permalink
report
parent
reply
44 points

Thanks admins, for taking this seriously.

Take care of yourselves as well. The images sound horrific.

permalink
report
reply
32 points
*

I think lemmy should allow the moderators to allow only text as posts in their communities, this way people would not be able to troll with images or videos in text-only communities.

It’s not the best solution, but it’s perfect for AskLemmy.

Maybe this website could use karma as a way to allow people to post images and/or videos. 100 general karma for pictures, 200 for videos, I don’t know.

It’s just a matter of being creative to avoid this kind of content.

permalink
report
reply
19 points
*

There are lots of ideas from people, but unfortunately Lemmy software doesn’t actually support most of them. And in terms of adding support to Lemmy, well there are just two devs plus community support so if people have the skills they may want to consider contributing code to implement some of these things.

Edit: just adding that some instances are building bots (kinda like an automod) but with 1,000+ instances we kinda need something built in

permalink
report
parent
reply
11 points
*

I think that’s at least partially why that Sublinks fork is starting. It’s in Java so the number of people capable of helping is much, much larger.

And while I’m excited to see how that turns out, I’ve got some reservations, particularly about some of the “moderation tools” being suggested, but if it breathes new life and excitement into a fediverse Reddit replacement, that’s a good thing.

permalink
report
parent
reply
10 points

The cool thing about ActivityPub is that we don’t have to pick one. Some instances run Kbin, some Mbin, some Lemmy. Some can run Sublinks and everyone gets to interact with each other.

The way federation works, instances can’t force their stuff on other instances. If a post on a community on your instance is removed by an admin of a different instance, then you’ll still see it on yours and they won’t see it on theirs (the exception is if the post is in a community where a moderator of that community or admin of that instance removes it, then it removes it for everyone - though Lemmy still has some quirks in regards to that), so different instances can have different moderation policies and you join one that matches the moderation policy you want.

permalink
report
parent
reply
4 points
*

If people want to help they can contribute to lemmy directly, or write moderation tools in other languages. It won’t help anyone to spend 8+ man years of development only to reach feature parity.

permalink
report
parent
reply
5 points

Especially make newer accounts text only, then there should be a new idea:

Image posts should not be federated for hours, maybe even a day. That would limit the spread, and provide local mods time to detect and remove content.

permalink
report
parent
reply
31 points

Well thanks for the spoiler thing, but I don’t even know what the acronym (is it even an acronym?) means anyway and now I’m too afraid to do a web search for it 😅

Well, maybe it’s better that way.

permalink
report
reply
32 points
*

It’s safe to look things up!

Looking up the name of a crime does not mean that you’re doing that crime.

If you look up “bank robbery” that doesn’t make you guilty of bank robbery. It doesn’t even mean you’re trying to rob a bank, or even want to rob a bank. You could want to know how bank robbers work. You could be interested in being a bank guard or security engineer. You could be thinking of writing a heist story. You could want to know how safe your money is in a bank: do they get robbed all the time, or not?

Please, folks, don’t be afraid to look up words. That’s how you learn stuff.

permalink
report
parent
reply
-6 points

To be fair… Also how you end up on a list.

permalink
report
parent
reply
21 points
*

::: spoiler CSAM is child sexual abuse material I believe. So yeah, better not to look it up :::

permalink
report
parent
reply
24 points

I’m pretty convinced the initialism was created so that people could Google it in an academic context without The Watchers thinking they were looking for the actual content.

permalink
report
parent
reply
6 points

You may be correct although it seems like pretty dumb reasoning. I doubt any of those cretins would search the words “child sexual abuse material.” That would require acknowledging the abuse part of it.

permalink
report
parent
reply
4 points

The initialism was created to focus the efforts of law enforcement. They have limited resources, so they want to address actual children being abused, rather than Japanese cartoons. Both are child porn, but CSAM involves real children.

permalink
report
parent
reply
5 points

Your spoiler didn’t work, apparently you need to write spoiler twice.

permalink
report
parent
reply
7 points

The second “spoiler” is actually text for the spoiler.

This is customisable text
::: spoiler This is customisable text 
permalink
report
parent
reply
4 points

Works in my app :P I really wish stuff like this were more standardized across the platform. Not really much point in spoilering it now since everyone is chatting about it.

permalink
report
parent
reply
2 points

thanks for the info!

permalink
report
parent
reply
5 points

I’ll just say illegal content involving minors.

permalink
report
parent
reply
23 points

This is exactly what happened in various other “alt-reddit” sites that I was a part of. CSAM is uploaded as a tactic so that they can then contact the registrar of said website and claim that this site is hosting said CSAM.

It’s basically a sign of the site becoming more popular and pissing off the larger players.

permalink
report
reply
7 points

Yeah we got some request like that and even for a announcement of ours. But we are planning to reduce them.

permalink
report
parent
reply
-4 points

But how would they even get the CSAM in the first place? It’s not like everyone just has pictures of children being sexually abused?

permalink
report
parent
reply
7 points

We aren’t talking about upstanding citizens here

permalink
report
parent
reply

Lemmy.World Announcements

!lemmyworld@lemmy.world

Create post

This Community is intended for posts about the Lemmy.world server by the admins.

Follow us for server news 🐘

Outages 🔥

https://status.lemmy.world

For support with issues at Lemmy.world, go to the Lemmy.world Support community.

Support e-mail

Any support requests are best sent to info@lemmy.world e-mail.

Report contact

Donations 💗

If you would like to make a donation to support the cost of running this platform, please do so at the following donation URLs.

If you can, please use / switch to Ko-Fi, it has the lowest fees for us

Join the team

Community stats

  • 1.5K

    Monthly active users

  • 800

    Posts

  • 36K

    Comments