this post was submitted on 21 Aug 2025
185 points (89.0% liked)

Selfhosted

50688 readers
525 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 2 years ago
MODERATORS
 

Some thoughts on how useful Anubis really is. Combined with comments I read elsewhere about scrapers starting to solve the challenges, I'm afraid Anubis will be outdated soon and we need something else.

you are viewing a single comment's thread
view the rest of the comments
[–] rtxn@lemmy.world 188 points 1 day ago* (last edited 1 day ago) (4 children)

The current version of Anubis was made as a quick "good enough" solution to an emergency. The article is very enthusiastic about explaining why it shouldn't work, but completely glosses over the fact that it has worked, at least to an extent where deploying it and maybe inconveniencing some users is preferable to having the entire web server choked out by a flood of indiscriminate scraper requests.

The purpose is to reduce the flood to a manageable level, not to block every single scraper request.

[–] 0_o7@lemmy.dbzer0.com 16 points 15 hours ago

The article is very enthusiastic about explaining why it shouldn't work, but completely glosses over the fact that it has worked

This post was originally written for ycombinator "Hacker" News which is vehemently against people hacking things together for greater good, and more importantly for free.

It's more of a corporate PR release site and if you aren't known by the "community", calling out solutions they can't profit off of brings all the tech-bros to the yard for engagement.

[–] loudwhisper@infosec.pub 3 points 13 hours ago

Exactly my thoughts too. Lots of theory about why it won't work, but not looking at the fact that if people use it, maybe it does work, and when it won't work, they will stop using it.

[–] poVoq@slrpnk.net 89 points 1 day ago* (last edited 1 day ago) (4 children)

And it was/is for sure the lesser evil compared to what most others did: put the site behind Cloudflare.

I feel people that complain about Anubis have never had their server overheat and shut down on an almost daily basis because of AI scrapers 🤦

[–] daniskarma@lemmy.dbzer0.com -3 points 14 hours ago (1 children)

I still think captchas are a better solution.

In order to surpass them they have to run AI inference which is also comes with compute costs. But for legitimate users you don't run unauthorized intensive tasks on their hardware.

[–] poVoq@slrpnk.net 7 points 12 hours ago (1 children)

They are much worse for accessibility, and also take longer to solve and are more distruptive for the majority of users.

[–] daniskarma@lemmy.dbzer0.com -1 points 12 hours ago* (last edited 12 hours ago) (1 children)

Anubis is worse for privacy. As you have to have JavaScript enabled. And worse for the environment as the cryptographic challenges with PoW are just a waste.

Also reCaptcha types are not really that disturbing most of the time.

As I said, the polite thing you just be giving users the options. Anubis PoW running directly just for entering a website is one of the most rudest piece of software I've seen lately. They should be more polite, and just give an option to the user, maybe the user could chose to solve a captcha or run Anubis PoW, or even just having Anubis but after a button the user could click.

I don't think is good practice to run that type of software just for entering a website. If that tendency were to grow browsers would need to adapt and straight up block that behavior. Like only allow access to some client resources after an user action.

[–] poVoq@slrpnk.net 9 points 12 hours ago* (last edited 12 hours ago) (1 children)

Are you seriously complaining about an (entirely false) negative privacy aspect of Anubis and then suggest reCaptcha from Google is better?

Look, no one thinks Anubis is great, but often it is that or the website becoming entirely inaccessible because it is DDOSed to death by the AI scrapers.

[–] daniskarma@lemmy.dbzer0.com -1 points 12 hours ago (1 children)

First, I said reCaptcha types, meaning captchas of the style of reCaptcha. That could be implemented outside a google environment. Secondly, I never said that types were better for privacy. I just said Anubis is bad for privacy. Traditional captchas that work without JavaScript would be the privacy friendly way.

Third, it's not a false proposition. Disabling JavaScript can protect your privacy a great deal. A lot of tracking is done through JavaScript.

Last, that's just the Anubis PR slogan. Not the truth, as I said ddos mitigation could be implemented in other ways. More polite and/or environmental friendly.

Are you astrosurfing for anubis? Because I really cannot understand why something as simple as a landing page with a button "run PoW challenge" would be that bad

[–] poVoq@slrpnk.net 3 points 12 hours ago (1 children)

Anubis is not bad for privacy, but rather the opposite. Server admins explicitly chose it over commonly available alternatives to preserve the privacy of their visitors.

If you don't like random Javascript execution, just install an allow-list extension in your browser 🤷

And no, it is not a PR slogan, it is the live experience of thousands of server admins (me included) that have been fighting with this for month now and are very grateful that Anubis has provided some (likely only temporary) relief from that.

And I don't get what the point of an extra button would be when the result is exactly the same 🤷

[–] grysbok@lemmy.sdf.org 4 points 10 hours ago

Latest version of Anubis has a JavaScript-free verification system. It isn't as accurate, so I allow js-free visits only if the site isn't being hammered. Which, tbf, prior to Anubis no one was getting in, JS or no JS.

[–] tofu@lemmy.nocturnal.garden 16 points 1 day ago* (last edited 1 day ago) (1 children)

Yeah, I'm just wondering what's going to follow. I just hope everything isn't going to need to go behind an authwall.

[–] rtxn@lemmy.world 34 points 1 day ago (2 children)
[–] grysbok@lemmy.sdf.org 18 points 1 day ago (1 children)

I'll say the developer is also very responsive. They're (ambiguous 'they', not sure of pronouns) active in a libraries-fighting-bots slack channel I'm on. Libraries have been hit hard by the bots: we have hoards of tasty archives and we don't have money to throw resources at the problem.

[–] lilith267@lemmy.blahaj.zone 7 points 1 day ago (1 children)

The Anubis repo has an enbyware emblem fun fact :D

[–] grysbok@lemmy.sdf.org 4 points 1 day ago

Yay! I won't edit my comment (so your comment will make sense) but I checked and they also list they/them on their github profile

[–] tofu@lemmy.nocturnal.garden 8 points 1 day ago

Cool, thanks for posting! Also the reasoning for the image is cool.

[–] mobotsar@sh.itjust.works 4 points 1 day ago (2 children)

Is there a reason other than avoiding infrastructure centralization not to put a web server behind cloudflare?

[–] poVoq@slrpnk.net 19 points 1 day ago

Yes, because Cloudflare routinely blocks entire IP ranges and puts people into endless captcha loops. And it snoops on all traffic and collects a lot of metadata about all your site visitors. And if you let them terminate TLS they will even analyse the passwords that people use to log into the services you run. It's basically a huge survelliance dragnet and probably a front for the NSA.

[–] bjoern_tantau@swg-empire.de 9 points 1 day ago (1 children)

Cloudflare would need https keys so they could read all the content you worked so hard to encrypt. If I wanted to do bad shit I would apply at Cloudflare.

[–] mobotsar@sh.itjust.works 7 points 1 day ago* (last edited 1 day ago) (2 children)

Maybe I'm misunderstanding what "behind cloudflare" means in this context, but I have a couple of my sites proxied through cloudflare, and they definitely don't have my keys.

I wouldn't think using a cloudflare captcha would require such a thing either.

[–] starkzarn@infosec.pub 12 points 1 day ago (1 children)

That's because they just terminate TLS at their end. Your DNS record is "poisoned" by the orange cloud and their infrastructure answers for you. They happen to have a trusted root CA so they just present one of their own certificates with a SAN that matches your domain and your browser trusts it. Bingo, TLS termination at CF servers. They have it in cleartext then and just re-encrypt it with your origin server if you enforce TLS, but at that point it's meaningless.

[–] mobotsar@sh.itjust.works 7 points 22 hours ago

Oh, I didn't think about the fact that they're a CA. That's a good point; thanks for the info.

[–] bjoern_tantau@swg-empire.de 4 points 1 day ago* (last edited 1 day ago)

Hmm, I should look up how that works.

Edit: https://developers.cloudflare.com/ssl/origin-configuration/ssl-modes/#custom-ssltls

They don't need your keys because they have their own CA. No way I'd use them.

Edit 2: And with their own DNS they could easily route any address through their own servers if they wanted to, without anyone noticing. They are entirely too powerful. Is there some way to prevent this?

[–] AnUnusualRelic@lemmy.world 19 points 1 day ago (1 children)

The problem is that the purpose of Anubis was to make crawling more computationally expensive and that crawlers are apparently increasingly prepared to accept that additional cost. One option would be to pile some required cycles on top of what's currently asked, but it's a balancing act before it starts to really be an annoyance for the meat popsicle users.

[–] rtxn@lemmy.world 23 points 1 day ago

That's why the developer is working on a better detection mechanism. https://xeiaso.net/blog/2025/avoiding-becoming-peg-dependency/