this post was submitted on 23 Dec 2025
73 points (94.0% liked)

Selfhosted

53934 readers
313 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

  7. No low-effort posts. This is subjective and will largely be determined by the community member reports.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 2 years ago
MODERATORS
 

I'm using CloudFlare to hide my home IP and to reduce traffic from clankers. However, I'm using the free tier, so how am I the product? What am I sacrificing? Is there another way to do the above without selling my digital soul?

you are viewing a single comment's thread
view the rest of the comments
[–] ptz@dubvee.org 5 points 4 days ago* (last edited 4 days ago) (1 children)

I've got bot detection setup in Nginx on my VPS which used to return 444 (Nginx for "close the connection and waste no more resources processing it), but I recently started piping that traffic to Nepenthes to return gibberish data for them to train on.

I documented a rough guide in the comment here. Of relevance to you are the two .conf files at the bottom. In the deny-disallowed.conf, change the line for return 301 ... to return 444

I also utilize firewall and fail2ban in the VPS to block bad actors, overly-aggressive scrapers, password brute forces, etc and the link between the VPS and my homelab equipment never sees that traffic.

In the case of a DDoS, I've done the following:

  • Enable aggressive rate limits in Nginx (it may be slow for everyone but it's still up)
  • Just stop either Wireguard or Nginx on the VPS until the storm blows over. (Crude but useful to avoid any bandwidth overages if you're charged for inbound traffic).

Granted, I'm not running anything mission-critical, just some services for friends and family, so I can deal with a little downtime.

[–] mesamunefire@piefed.social 4 points 4 days ago (1 children)

I have something similar with fail2ban + hidden buttons. If the requester goes and clicks on the hidden buttons on the main site, it gets into a rabbit hole. After 3 requests, it gets banned for a bit. Usually stops the worst offenders. OpenAI and some of the scrapers are the worst.

Google/bing, I do actually see them hit robots.txt then jump off, which is what they should be going.

[–] ptz@dubvee.org 1 points 4 days ago* (last edited 4 days ago) (1 children)

Oooooh. That's smart. I mostly host apps, but in theory, I should be able to dynamically modify the response body and tack on some HTML for a hidden button and do that.

I used to disallow everything in robots.txt but the worst crawlers just ignored it. Now my robots.txt says all are welcome and every bot gets shunted to the tarpit 😈

[–] mesamunefire@piefed.social 1 points 4 days ago (1 children)

Nice! Thats another way to do it. 😀

I know others use Arabis(?) I think thats what it called. The anime girl one that does a calc on top. Ive never had good luck with it. I think bot are using something to get around and it messes with my requests. Might also be my own fiddling.

[–] FrostyPolicy@suppo.fi 2 points 4 days ago (1 children)

I know others use Arabis(?) I think thats what it called.

You probably mean Anubis.