I like to code, garden and tinker

  • 0 Posts
  • 20 Comments
Joined 1 year ago
cake
Cake day: February 9th, 2024

help-circle
  • Centralization is a weakness. These services can be targeted by governments that want to limit communication. Free speech is a commodity, and servers host this free speech. If a hostile organization, such as a government, targets a channel of free speech such as those hosted on a platform that makes it easy to setup a mastodon instance, this become an easy target that will affect a large portion of users. If you are serious about freedom, you have the freedom to self-host your own platforms.

    Edit: I realize my post doesn’t answer the question proposed, but it’s more of an argument against such services. I would argue self-hosting doesn’t rely on paying third-parties to host your software, but I guess this is in the eye of the hoster.



  • dudeami0@lemmy.dudeami.wintoSelfhosted@lemmy.worldNon-Cloudflare AI blocking?
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    1
    ·
    4 months ago

    The only way I can think of is require users to authenticate themselves, but this isn’t much of a hurdle.

    To get into the details of it, what do you define as an AI bot? Are you worried about scrappers grabbing the contents of you website? What is the activities of an “AI Bot”. Are you worried about AI bots registering and using your platform?

    The real answer is not even cloudflare will fully defend you from this. If anything cloudflare is just making sure they get paid for access to your website by AI scappers. As someone who has worked around bot protections (albeit in a different context than web scrapping), it’s a game of cat and mouse. If you or some company you hire are not actively working against automated access, you lose as the other side is active.

    Just think of your point that they are using residential IP addresses. How do they get these addresses? They provide addons/extensions for browsers that offer some service (generally free VPNs) in exchange for access to your PC and therefore your internet in the contract you agree to. The same can be used by any addon, and if the addon has permissions to read any website they can scrape those websites using legit users for whatever purposes they want. The recent exposure of the Honey scam highlights this, as it’s very easy to get users to install addons by selling users they might save a small amount of money (or make money for other programs). There will be users who are compromised by addons/extensions or even just viruses that will be able to extract the data you are trying to protect.



  • My question would be, why do you need a more powerful server? Are you monitoring your load and seeing it’s overloaded often? Are you just looking to be able to hook more drives to it? Do you need to re-encode video on the fly for other devices? Giving some more details would help someone to give a more insightful answer. I personally am using a Raspberry Pi 4, Chromebox w/ an i7, an old HP rack server, and an old desktop PC for my self hosting needs, as this is cheaper than buying all new hardware (though the electricity bill isn’t the greatest haha, but oh well). If you are just looking for more storage, using the USB 3.0 slots on the Raspberry Pi 4b you can add a couple extra SSDs using a NVMe to USB 3.0 enclosure. For most purposes the speeds will be fine for most applications.

    As for SSD vs HDD, SSD hands down. The only reason you’d pick an HDD is if your trying to get more storage cheaper and don’t mind a higher rate of failure. If your data is at all valuable, and it almost always is, redundancy should be added as well.

    And as for running Linux, if it can’t run Linux I wouldn’t want to own it.

    Edit: Fixed typo


  • This might help, sorry if it doesn’t, but here is a link to CloudFlares 5xx error code page on error 521. If you’ve done everything in the resolution list your ISP might be actively blocking you from hosting websites, as it is generally against the ISPs ToS to do such on residential service lines. This is why I personally rent a VPS and have a wireguard VPN setup to host from the VPN, which is basically just a roll your own version of Tailscale using any VPS provider. This way you don’t need to expose anything via your ISPs router/WAN and they can’t see what you are sending or which ports you are sending on (other than the encrypted VPN traffic to your VPS of course).








  • I chose Vultr because a friend recommended it, and the generous egress bandwidth provided. With a single $5 VPS you get 2TB free egress (this is for the whole account) + 1TB earned over the month (it gives the 1TB evenly out as the month progresses). This is more bandwidth than I’ll probably ever use for my services so I don’t have to be too concerned about bandwidth usage. So far I’ve hosted a few game servers and some HTTP web services (including my lemmy instance) with no issue.



  • If you are seeing your routers config page, and you are sure you are connecting from outside your network, it sounds like the router’s 443 page is overriding the port forwarding. Otherwise, it’s like @fixmycode@feddit.cl said and you just need a local DNS that points to the right spot locally, and let your public DNS point for external connections.

    As for the hosts file, you can see a guide here for windows/linux/mac. Basically this is a override of any DNS entries. Here you can point jellyfin.domain.com to your jellyfin servers LAN IP and test the connection works.


  • To me this seems like a routing issue. Some things to check:

    • Can a port checking service (like canyouseeme) see 443?
    • Can you connect to 443 from the internet (use a VPN or separate network)?
    • Can you edit your hosts file to point jellyfin.mydomain.com to your jellyfin server’s LAN IP?
    • After doing this, can you connect to jellyfin.mydomain.com on the LAN?

    My guess is the router is routing traffic to it’s external IP from the LAN back to itself, without following port forwarding rules. Good luck figuring it out though!



  • I’m not sure if there really is issues, I think it’s just new ground since most lemmy instances have been able to run on a single node due to the low populations. It seems most large public instances are just adding bigger servers to deal with the problem short term.

    From what I can tell (I am not an expert in this field), it seems most of the architecture would spread horizontally without much issue. I haven’t seen anywhere this is done yet, but I could be missing the obvious.

    The lemmy backend api just takes HTTP requests (and at the present websockets, but this is changing in 0.18 to only HTTP requests), and it uses postgres as the backend storage. Using a kubernetes postgres operator to scale the database and then running multiple lemmy backend api instances (and frontend as needed) seems like it would work, or would require minimal work to get running.



  • I currently host a few services (including the lemmy instance I am replying from) behind a commodity $5 VPS, while the services are actually hosted locally. I setup WireGuard to have a simple hard-coded peer-to-peer VPN connection from my local client to the remote VPS, and then setup some iptable rules on the VPS to redirect traffic to the VPN network. This allows me to host behind a NAT (my biggest issue), but also handles IP changes and hides your home’s public IP. I am no networking engineer, so I am not sure how safe this is, manually routing packets can be tricky.

    There are also a few services this does not work for. So far I’ve found CS:GO dedicated servers (if the public IP of the local machine differs from the VPS) and email servers cannot be behind a NAT to function properly. Other services likely exist, but you’ll be able to run most services. You do lose the originating IP addresses in this case, which can complicate things (the case for email servers).

    This process is explained in detail on wickedyoda.com and with a video tutorial.