Are there any protections to ensure that web scraping bots don’t overload the site? I was thinking about Anubis - GitHub - TecharoHQ/anubis: Weighs the soul of incoming HTTP requests to stop AI crawlers - and while I see that IF only accepts “browser connections”, there are many bots that ignore robots.txt and will pretend to be “real humans”
we already have the security system, which blocks most low-effort scrapers. and fun fact, anubis is even more heavier to host for millions of websites in free hosting. ![]()
I’m not sure if Anubis is actually harder to run for us, but I think that the system we have should be adequate while also being less intrusive than Anubis.
If you’re not confident in our solution and have your own subdomain, you could also setup Cloudflare for your domain, who also have their own anti-AI solution.
just code a simple php code that force user to solve captcha like cloudflare turnstile or google recaptcha, make it simple so that people dont suspect your site is a fake captcha scam website
This topic was automatically closed 7 days after the last reply. New replies are no longer allowed.