9 points

#TL;DR:

User-agent: GPTBot
Disallow: /
User-agent: ChatGPT-User
Disallow: /
User-agent: Google-Extended
Disallow: /
User-agent: PerplexityBot
Disallow: /
User-agent: Amazonbot
Disallow: /
User-agent: ClaudeBot
Disallow: /
User-agent: Omgilibot
Disallow: /
User-Agent: FacebookBot
Disallow: /
User-Agent: Applebot
Disallow: /
User-agent: anthropic-ai
Disallow: /
User-agent: Bytespider
Disallow: /
User-agent: Claude-Web
Disallow: /
User-agent: Diffbot
Disallow: /
User-agent: ImagesiftBot
Disallow: /
User-agent: Omgilibot
Disallow: /
User-agent: Omgili
Disallow: /
User-agent: YouBot
Disallow: /
permalink
report
reply
7 points

Of course, nothing stops a bot from picking a user agent field that exactly matches a web browser.

permalink
report
parent
reply
3 points

Nothing stops a bot from choosing to not read robots.txt

permalink
report
parent
reply
2 points
*

Indeed, as has already been said repeatedly in other comments.

permalink
report
parent
reply
111 points

Block? Nope, robots.txt does not block the bots. It’s just a text file that says: “Hey robot X, please do not crawl my website. Thanks :>”

permalink
report
reply
13 points

Unfortunate indeed.

“Can AI bots ignore my robots.txt file? Well-established companies such as Google and OpenAI typically adhere to robots.txt protocols. But some poorly designed AI bots will ignore your robots.txt.”

permalink
report
parent
reply
23 points

typically adhere. but they don’t have to follow it.

poorly designed AI bots

Is it a poor design if its explicitly a design choice to ignore it entirely to scrape as much data as possible? Id argue its more AI bots designed to scrape everything regardless of robots.txt. That’s the intention. Asshole design vs poor design.

permalink
report
parent
reply
45 points

Robots.txt is honor-based and Big Data has no honor.

permalink
report
parent
reply
7 points

This is why I block in a htaccess:

# Bot Agent Block Rule
RewriteEngine On
RewriteCond %{HTTP_USER_AGENT} (BOTNAME|BOTNAME2|BOTNAME3) [NC]
RewriteRule (.*) - [F,L]
permalink
report
parent
reply
19 points

This is still relying on the bot being nice enough to tell you that it’s a bot; it could just not.

permalink
report
parent
reply
7 points

Exactly. The only truly effectively way I’ve ever found to block bots is to use a service like Akamai. They have an add-on called Bot Manager that identifies requests as bots in real time. They have a library of over 1000 known bots and can also identify unknown bots built on different frameworks, bots that impersonate well known bots like Googlebot, etc. This service is expensive, but effective…

permalink
report
parent
reply
60 points

I disallow a page in my robots.txt and ip-ban everyone who goes there. Thats pretty effective.

permalink
report
parent
reply
4 points

smart

permalink
report
parent
reply
9 points

Can you explain this more?

permalink
report
parent
reply
25 points

Imagine posting a rule that says “do not walk on the grass” among other rules and then banning anyone who steps on the grass with the thought process that if they didn’t obey that rule they were likely disobeying other rules. Except the grass is somewhere that no one would see unless they actually read the rules. The rules were the only place that mentioned that grass.

permalink
report
parent
reply
16 points

Did you ban it in your humans.txt too?

permalink
report
parent
reply
18 points
*

humans typically don’t visit [website]/fdfjsidfjsidojfi43j435345 when there’s no button that links to it

permalink
report
parent
reply
5 points

Not sure if that is effective at all. Why would a crawler check the robots.txt if it’s programmed to ignore it anyways?

permalink
report
parent
reply
16 points

cause many crawlers seem to explicitly crawl “forbidden” sites

permalink
report
parent
reply
3 points

Google and script kiddies copying code…

permalink
report
parent
reply
1 point

You could also place the same page as a hidden link on your home page.

permalink
report
parent
reply
4 points

I doubt it’d be possible in most any way due to lack of server control, but I’m definitely gonna have to look this up to see if anything similar could be done on a neocities site.

permalink
report
parent
reply
7 points

Is the page linked in the site anywhere, or just mentioned in the robots.txt file?

permalink
report
parent
reply
10 points

Only in the robots.txt

permalink
report
parent
reply
2 points

Can this be done without fail2ban?

permalink
report
parent
reply
1 point

probably. never used it tho

permalink
report
parent
reply
1 point

Should be able to do it with Crowdsec

permalink
report
parent
reply
10 points

It isn’t an enforceable solution. robots.txt and similar are just please bots dont index these pages. Doesn’t mean any bots will respect it

permalink
report
reply
24 points

This does not block anything at all.

It’s a 1994 “standard” that requires voluntary compliance and the user-agent is a string set by the operator of the tool used to access your site.

https://en.m.wikipedia.org/wiki/Robots.txt

https://en.m.wikipedia.org/wiki/User-Agent_header

In other words, the bot operator can ignore your robots.txt file and if you check your webserver logs, they can set their user-agent to whatever they like, so you cannot tell if they are ignoring you.

permalink
report
reply
19 points

robots.txt will not block a bad bot, but you can use it to lure the bad bots into a “bot-trap” so you can ban them in an automated fashion.

permalink
report
reply
9 points

I’m guessing something like:

Robots.txt: Do not index this particular area.

Main page: invisible link to particular area at top of page, with alt text of “don’t follow this, it’s just a bot trap” for screen readers and such.

Result: any access to said particular area equals insta-ban for that IP. Maybe just for 24 hours so nosy humans can get back to enjoying your site.

permalink
report
parent
reply
2 points

Problem is that you’re also blocking search engines to index your site, no?

permalink
report
parent
reply
8 points

Nope. Search engines should follow the robots.txt

permalink
report
parent
reply
2 points

Not if they obeyed the rules

permalink
report
parent
reply
5 points
*

Robots.txt: Do not index this particular area.

Problem is that you’re also blocking search engines to index your site, no?

No. That’s why they wrote “this particular area”.

The point is to have an area of the site that serves no purpose other than to catch bots that ignore the rules in robots.txt. Legit search engine indexers will respect directives in robots.txt to avoid that area; they will still index everything else. Bad bots will ignore the directives, index the forbidden area anyway, and by doing so, reveal themselves in the server logs.

That’s the trap, aka honeypot.

permalink
report
parent
reply

Privacy

!privacy@lemmy.ml

Create post

A place to discuss privacy and freedom in the digital world.

Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.

In this community everyone is welcome to post links and discuss topics related to privacy.

Some Rules

  • Posting a link to a website containing tracking isn’t great, if contents of the website are behind a paywall maybe copy them into the post
  • Don’t promote proprietary software
  • Try to keep things on topic
  • If you have a question, please try searching for previous discussions, maybe it has already been answered
  • Reposts are fine, but should have at least a couple of weeks in between so that the post can reach a new audience
  • Be nice :)

Related communities

Chat rooms

much thanks to @gary_host_laptop for the logo design :)

Community stats

  • 3.9K

    Monthly active users

  • 3K

    Posts

  • 78K

    Comments