The steady stream of people who are telling me that the Santa moderation bot is going to delete anyone who’s downvoted or disagrees with the group, is continuing unabated.

Here’s an olive branch: You’ve got a point. It’s just a black box and I juggle the parameters to some secret process to ban the people who got some downvotes, I can understand how that comes across as toxic. I might or might not be lying about taking careful time to look over its judgements and make sure that I think the impact is more positive than negative, but at the end of the day, it doesn’t matter. You still have to trust my intentions and trust the bot to make good decisions, and trusting that to an automated system rarely works out well.

To me, delegating the moderation of the community to the segment of that community that’s trusted and consistently upvoted by the rest of us is better than giving it to a handful of people who wield unilateral power according to random rules. I like the bot’s judgements most of the time when I look at them. The question is simply whether this algorithm is actually doing that delegation effectively, or if it’s just banhammering anyone who gets a couple of downvotes. I’m confident that it’s doing the first thing almost all of the time.

In talks behind the scenes with other moderators, I’ve been going into a lot of detail about specific users and going back and forth about judgements. I also do a ton of checking behind the scenes. I don’t want to do that publicly. I think it would be deeply informative to post a list of the “top ten” and “bottom ten” users, and go into detail about why the low-ranked users got where they are, but that’s probably not a good idea.

What I would like to do is share that information on some level, so that people can see what’s going on, instead of it being me relaying that everything’s good. It’s tough because I can’t break down every level of detail without invading all kinds of people’s privacy. That said, I do think that there’s a way to be found to open up the process so people can see and give input to what’s going on.

One happy medium I could do would be to have the bot post its spot-check automatically about once a week. It could pick out one random user who’s barely on the borderline, and post a couple of the worst comments they made. Usually, when I’m messing around with its parameters, that’s what I am trying to do. There are some comments that are clearly toxicity that have no business anywhere. There are some comments that are clearly free speech, and even if they’re getting downvotes, they deserve to be heard. Then there are some comments that are on the borderline between. My goal is to set up the parameters so that the borderline rank value for a ban matches up with the users who are on that borderline.

I can see some upsides and downsides to posting that publicly. What do people think, though? What would you want to see, in order to make an informed decision about what you think of this whole approach?

3 points
*

I think you’re right to be concerned with the trade-off between transparency and privacy. ActivityPub infrastructure technically exposes a lot of things that would be only shared between company employees and their advertising partners in other social media, but due to the discretion of the people implementing front-end software, most of that information is not exposed to the general public. While the Fediverse has technical transparency, it also has functional privacy. The developers of Lemmy frontends deserve a lot of credit for their caution and care.

I think a spot-check on a member of the Fediverse would make the inner workings of SantaBot easier to understand. It may be difficult to do that though, without breaking some of the norms about member privacy that we have been carefully building as a culture.

One solution is to allow members to opt-in to allow their Santabot analysis to be shared publicly. I think I might be one of your borderline cases; I give my consent if you’d like to use me as an example.

permalink
report
reply
-157 points

I agree. I think spot-checking can do a lot to bring transparency into the picture, and if it’s done carefully, then it’ll be possible to avoid exposing too much about people who haven’t agreed to have it exposed about them.

I thought about it for a while, and I think doing a weekly spot-check post for a handful of controversial users, showing a visualization of their rank and where it is coming from, might work. Here’s one quickly hacked-up example in the form of a bar code. Time goes from left to right, blue stripes are positive rank, and red stripes are negative rank. Here’s your breakdown for the last month:

There are three big red stripes. From left to right, they are these threads:

There is also plenty of blue, though, so you’re comfortably over the line as a nice person under the current parameter set. It’s worth mentioning that a lot of the blue stripes are “unpopular” opinions from the point of view of the average liberal, that are popular on Lemmy, or detailed takedowns of MBFC:

My opinion is that most of the time, someone who’s garnering a healthy mixture of blue and red is probably showing good faith, and when someone is managing to garner mostly red, it’s more likely to be an issue of quality of engagement, not even necessarily that they’re trying to say something unpopular that the bot is then censoring. But, of course, the proof is in how it works in practice on real users and real content.

I think doing some type of visualization, maybe automatically generated, and showing the progression over time of someone’s rank depending on particular comments, can help to inform the discussion. I’m sure it won’t stop people from accusing me of all kinds of malfeasance in the way the bot operates, but it can help to put more eyes on it from people who are open and interested in seeing how it’s working.

permalink
report
parent
reply
2 points
*

I love this – Reddit used to do a yearly thing where they’d send you your top upvoted and downvoted posts and comments that was always nostalgic and fascinating to me as a user. Like canvas, I think it’s an idea worth copying with a more federated framework.

Maybe you could write an action that allows Fediverse members to get a similar breakdown and visualization automatically generated and then delivered to them via direct message. People who are curious about how the bot works can message the bot and see how it views them, and then they can share the details publicly if they so choose. I think this could be really popular.

permalink
report
parent
reply
1 point

How about this?

That’s 30 days of Santa’s ranking for your user, showing the comment threads that made big impacts up or down. The dotted horizontal line is 0, and the cutoff for banning a person is down below that line. Here are some anonymized examples of people who got banned:

They were doing well until, in the pink part, they posted 28 comments heatedly insisting that there’s no genocide in Gaza.

I think this is informative about how the system works without being useful for gathering analytics to rig the system. You can see what kind of participation impacts it in what ways, and how to put it into the context of the sum total of your participation for the month, but the emphasis is on the comments and behavior instead of on the math. What do you think?

permalink
report
parent
reply
1 point

Is there a cap to how much stomp a user can have through their votes? By accumulating enough zeitgeist points, can a single user ban a new user from !pleasantpolitics with downvotes?

permalink
report
reply
2 points

There’s not a cap. That type of activity is, in fact, a classical failure mode of this type of network. Just like people learned to build link farms to artificially give page rank to fake pages, people can learn to farm for zeitgeist points to then give or take away rank from some targeted user. That is one reason I’m being cagey about giving away introspection tools or detailed road maps of people’s points. I don’t want to facilitate someone getting feedback about how well an effort like that is working.

I’m a little more concerned about people accumulating points and then upvoting a troll account to make sure it doesn’t get banned, than I am about people downvote-bombing someone they disagree with to ban them. They are both concerns, though. There are ways around both through tweaks to the algorithm, but I’ve constantly been surprised about how the tools work out in practice as compared to my theory about them, and so I’m waiting until it happens before I start messing with solutions to it. I do have some ideas in mind for how to deal with it. I am guessing that in the long run, it won’t be too big an issue, but I want to see how it works out in practice before actualizing the countermeasures I was thinking of.

permalink
report
parent
reply
2 points

I played around with possibilities for a while, and did some more fixing and tweaking of the algorithm and visualization tools. Here’s one way I think it could work. Once a week, the bot could post a breakdown of three random users who are permitted to post, and three random users who aren’t permitted to post. Right now, that breakdown would be:

Permitted to post:

Not permitted to post:

That means that anyone who wants to can check up on how it’s making its decisions. Then, in addition, anyone who wants an explanation for their user, I can do that.

Those charts are anonymized. I’ll send the users in question to some of the admins to see what they think. I think it’s okay to post a few users, as long as it’s random and not repetitive. I don’t think it would come across as singling anyone out or making them uncomfortable, but I’m curious what other people think.

permalink
report
reply
1 point
*

It might be fairly easy to de-anonymize users as not all users post in all threads, and identifying a user based on which threads they post in and generally what the response was to their posts isn’t impossible.

On the other hand, it doesn’t reveal information that we’ve decided should be treated special, like who is voting in the comments and posts. When posting a controversial twitter screenshot of a non-public figure, it’s internet etiquette and good form to blur the target’s name, even though the tweet can be found via text search. This ups the effort to attack the user a little, but also communicates through actions that trolling is being discouraged – which I think is the most effective deterrent.

The measures you’re taking seem to be in line with that internet etiquette. Especially considering the relatively small exposure your project is getting (at this point it seems it’s just us talking in this thread, for example) the precautions you have in place should be enough. You may consider revising this if you get complaints of harassment or when your project develops a much larger audience.

permalink
report
parent
reply

Santabot

!santabot@slrpnk.net

Create post

Santabot is an automated moderation tool, designed to reduce moderation load and remove bad actors in a way that is transparent to all and mostly resistant to abuse and evasion.

This is a community devoted to meta discussion for the bot.

Community stats

  • 330

    Monthly active users

  • 9

    Posts

  • 46

    Comments

Community moderators