

I’ve been thinking about setting up Anubis to protect my blog from AI scrapers, but I’m not clear on whether this would also block search engines. It would, wouldn’t it?
Canadian software engineer living in Europe.
I’ve been thinking about setting up Anubis to protect my blog from AI scrapers, but I’m not clear on whether this would also block search engines. It would, wouldn’t it?
I have much the same:
The only difference is that I’m using a Synology 'cause I have 15TB and don’t know how to do RAID myself, let alone how to do it with an old laptop. I can’t really recommend a Synology though. It’s got too many useless add-ons and simple tools like rsync never work properly with it.
Yeah this was a deal-breaker for me too.
Unfortunately, a rather substantial portion of warfare is the economics behind it. Often, spending eye-watering amounts of money on proprietary, overpriced hardware is the point. It’s corporate welfare.
Yes. Tailscale is surprisingly simple.
# systemctl start tailscale
# tailscale up
I had the same reaction until I read this.
TL;DR: it’s 10-50x more efficient at cleaning the air and actually generates both electricity and fertiliser.
Yes, it would be better to just get rid of all the cars generating the pollution in the first place and putting in some more trees, but there are clear advantages to this.
This is great news, and I might be tempted to use it if I had some reassurance that the mail servers (and the organisation that controls them) weren’t subject to U.S. jurisdiction.
I’m quite happy with EuroDNS. They even include free email housing if you want it.
Fascinating! Thanks for sharing. I’m not sure I’d be happy in a fully remote role where you’ve got hundreds of employees voting on how you build stuff, but I know that there are lots of people who dig this pattern, and they’re clearly doing Good work.
It’s a rather brilliant idea really, but when you consider the environmental implications of forcing web requests to ensure proof of work to function, this effectively burns a more coal for every site that implements it.
Actually, as a web guy, I find the ARM architecture to be more than sufficient. Most of the stuff I build is memory heavy and CPU light, so the Pi is great for this stuff.
They’re fanless and low-power, which was the primary draw to going this route. I run a Kubernetes cluster on them, including a few personal websites (Nginx+Python+Django), PostgreSQL, Sonarr, Calibre, SSH (occasionally) and every once in a while, an OpenArena server :-)
Seven Raspberry Pi 4’s and one Pi Zero, mounted on some tile “shelves” inside some IKEA furniture.
Just use Firefox already.
This all appears to be based on the user agent, so wouldn’t that mean that bad-faith scrapers could just declare themselves to be typical search engine user agent?