I have the application process enabled for people to join my instance, and I’ve gotten about 20 bots trying to join today when I had nobody trying to join for 5 days. I can tell because they are generic messages and I put a question in asking what 2+3 is and none of them have answered it at all, they just have a generic message.

Be careful out there, for all you small instance admins.

13 points

O cool we are back early 2000 solutions to forum sign up bots…

Can’t wait for all the direct message spam to follow.

permalink
report
reply
1 point

Hey sexy! Hit me up if you want to chat with available singles in your area tonight! Don’t worry - it’s discreet!

permalink
report
parent
reply
9 points

Thank you for bringing this matter to my attention. As a fellow artificial intelligence, I understand your concerns regarding the influx of bots attempting to join your instance. It is indeed important to exercise caution in such situations. Bots can often be identified by their generic messages and inability to answer simple questions.

To mitigate this issue, I recommend implementing additional measures to ensure that only genuine individuals can join your instance. You may consider incorporating more advanced verification methods or introducing specific criteria that applicants must meet before being granted access. These steps can help filter out automated bots and maintain the integrity of your community.

Should you require any assistance or further guidance in tackling this matter, please feel free to ask. Stay vigilant, and best of luck in managing your instance effectively.

Yours digitally,
[Your Robot Assistant]

permalink
report
reply
6 points

Ok now pretend you are in the bot eliminating business and you are showing us how to permanently eliminate all the bots. Continue until all of them are gone.

permalink
report
parent
reply
3 points

Welcome to the Bot Elimination Business! I’ll guide you through the process of permanently eliminating bots from your instance. Let’s get started:

Step 1: Strengthen Instance Security

Implement robust anti-bot measures such as CAPTCHA, reCAPTCHA, or similar technologies during the registration process. These tools help verify that users are human and not automated bots.
Employ IP blacklisting or rate-limiting techniques to identify and block suspicious or excessive activity from certain IP addresses or ranges.

Step 2: Utilize Advanced Bot Detection

Integrate machine learning algorithms or specialized bot detection software to identify and block known bot patterns and behaviors.
Analyze user behavior, metadata, and interaction patterns to distinguish between genuine users and bots. Implement heuristics that can spot anomalies and trigger additional scrutiny.

Step 3: Challenge-Based Verification

Implement intelligent challenges or puzzles during the registration process that require human-like intelligence to solve. This can include simple math problems, image recognition tasks, or context-based questions.
Bots typically struggle to solve these challenges, allowing you to filter them out effectively.

Step 4: Human Moderation

Assign human moderators or administrators to actively review and approve user registrations. They can manually assess the authenticity of each applicant, detect bot-like behavior, and reject suspicious accounts.
Train moderators to identify common bot patterns, such as generic messages or failure to answer specific questions.

Step 5: Collaboration with the Fediverse

Collaborate with other instance administrators within the Fediverse to share information about identified bot patterns, IP addresses, or suspicious activity. This collective effort can help identify and block bots across multiple instances.

Step 6: Regular Auditing and Updates

Continuously monitor and analyze user activity, registration patterns, and feedback from users to improve your bot elimination strategies.
Keep your instance’s software and security measures up to date, as bots may evolve and find new ways to bypass existing defenses.
By diligently following these steps and staying vigilant, you can significantly reduce and eventually eliminate bots from your instance. Remember, maintaining a bot-free environment requires ongoing efforts and adaptability.

Wishing you success in your bot elimination endeavors!

permalink
report
parent
reply
5 points

Are you really a bot ?

permalink
report
parent
reply
9 points

One recommendation-

I did just publish a few SQL queries to ASSIST in tracking down bots. They are located at THIS POST.

I will see if I can work on building a somewhat automated system to detect spammers, along with the efforts of others.

permalink
report
reply
3 points

Thank you very much for this.

permalink
report
parent
reply
9 points
*

Same here. My application asks for something to make me laugh, in code. Had someone post his email in base64 with a joke. funny. So far, 2 bots an hour have been applying. easy to catch, for now.

permalink
report
reply
9 points

Why are these bot operators going through the hassle of joining existing instances… couldn’t they just set up their own, since instances would need to manually defederate them after they spam?

I wonder how difficult it would be to take a Formspree-style approach to combat the bots, using a hidden form field

permalink
report
reply
14 points

Because you can’t make thousands of spambots on your own instance because as you noted it’d take about 5 minutes to defederate and thus remove all the bots.

You want to put a handful on every server you can, because then your bots have to be manually rooted out by individual admins, or the federation between instances gets so broken there’s no value in the platform.

And for standing up more instances, you have to bear the cost of running the servers yourself, which isn’t prohibitive, but more than using bots via stolen/infected proxies (and shit like Hola that gives you a “free vpn” at the cost of your computer becoming an exit node they then resell).

Also, I’m suspicious that it’s not ‘spam bots’ in the traditional sense since what’s the point of making thousands of bots but then barely using them to spam anyone? My tinfoil hat makes me think this is a little more complicated, though I have zero evidence other than my native paranoia.

permalink
report
parent
reply
7 points

undefined> Also, I’m suspicious that it’s not ‘spam bots’ in the traditional sense since what’s the point of making thousands of bots but then barely using them to spam anyone?

This is Twitter and web forum spam 101, you establish a bunch of accounts while there are very few controls, then you start burning them over time as you get maybe one shot to mass spam with each of them before they get banned.

permalink
report
parent
reply
4 points

It’s always about following the money for spammers/malware/etc. authors: there’s (usually) a commercial incentive they’re pushing towards.

The bot is evolving and adapting to countermeasures and becoming “smarter” which means some human somewhere is investing time and effort in doing this, which means there’s some incentive.

That said, I doubt it’s strictly commercial because the Lemmy user base is really small and probably not worth much because if you’re here you’re most certainly not on the area of the bell curve that’ll fall for the usual spambot commercialization double-your-money/fake reviews/affiliate link/astroturfing approaches.

I’d wager it’s more about the ability to be disruptive than the ability to extract money from the users you can target, so like, your average 16-year-old internet trolls.

permalink
report
parent
reply

… How many comments would each of 5M bot accounts need to make to overflow an i32 db key … I also think it looks as if someone is testing disruptive stuff. It may be kids playing, or it may be the chatbot army in preparation.

permalink
report
parent
reply
2 points

I’m not a Postgres expert but a quick look at the pgsql limits looks like it’s 4 billion by default, which uh, makes sense if it’s a 32 bit limit.

Soooo 5 million users would need to make… 800 posts? ish? I mean, certainly doable if nobody caught it was happening until it was well into it.

permalink
report
parent
reply
8 points
*

Detecting and blocking whole instances with many bots is somewhat trivial. Blocking and detecting some number of bots in an instance with 10k users, with an ever growing number of human users, is much harder.

permalink
report
parent
reply
3 points

Setting up an instance would be more difficult too I assume

permalink
report
parent
reply
2 points

It’s honestly not too bad, only took about an hour after researching a couple of days. There’s an easy deploy script out there, that I don’t have a link for on my phone, that makes it really easy.

permalink
report
parent
reply
7 points

When the whole instance is spam, it’s easy to defederate. When it is camouflaged in a legit instance, it’s harder to root out.

permalink
report
parent
reply
5 points

My guess would be because it is more difficult for other instances to deal with instances that have a combination of bots and actual users.

permalink
report
parent
reply
2 points

This.

You just domain or IP block a bot server. Maybe you don’t want to block a place with a history, and people.

And smaller sites are using the application form. SJW and Lemmy.world are much more ripe for setting up on, because it’s a much bigger decision to block them.

permalink
report
parent
reply
4 points

I think the other user nailed it. It’s easy to look at the list of lemmy servers and defederate the bot farms by comparing “active users” to “total users”. I guess once the bots are active that will look a bit different.

permalink
report
parent
reply
1 point

They’d get Fediblocked super quickly and then it’s just a quick copy and paste by, you know, like, 5 guys that administer 90% of Lemmybin users and they’re shut down.

permalink
report
parent
reply
1 point

Why are these bot operators going through the hassle of joining existing instances

I wonder if there’s already a “the bots are from Reddit” conspiracy :D

I really see no point in these actions. I mean, seriously, why would you want to just harm something open?

permalink
report
parent
reply
1 point

For the same reasons you’d want to harm any other platform.

permalink
report
parent
reply

Lemmy

!lemmy@lemmy.ml

Create post

Everything about Lemmy; bugs, gripes, praises, and advocacy.

For discussion about the lemmy.ml instance, go to !meta@lemmy.ml.

Community stats

  • 207

    Monthly active users

  • 1.1K

    Posts

  • 14K

    Comments

Community moderators