You can use cloudflares DNS and not use their WAF (the proxy bit) just fine. I have been for almost a decade.
🇨🇦
You can use cloudflares DNS and not use their WAF (the proxy bit) just fine. I have been for almost a decade.
I’m not sure whether this is specific to this project, docker, or YAML in general.
Looking through my other 20 or so compose files, I use the array notation for most of my environment variables, but I don’t have any double quotation marks elsewhere. Maybe they’re not supposed to work in this format, idk.
Good to keep in mind I guess.
Dev replied to my github discussion.
Apparently it’s an issue with array style env variable layout.
environment:
key:"value"
Instead of
environment:
- key=value
Trying to set that up to try out, but I can’t get it to see/use my config.yaml.
/srv/filebrowser-new/data/config.yaml
volumes:
Says ‘/config/config.yaml’ doesn’t exist and will not start. Same thing if I mount the config file directly, instead of just its folder.
If I remove the env var, it changes to “could not open config file ‘config.yaml’, using default settings” and starts at least. From there I can ‘ls -l’ through docker exec and see that my config is mounted exactly where it’s supposed to be ‘/config/config.yaml’ and has 777 perms, but filebrowser insists it doesn’t exist…
My config is just the example for now.
I don’t understand what I could possibly be doing wrong.
/edit: three hours of messing around and I figured it out:
Must not have quotation marks. Removed them and now it’s working.
FolderSync selectively syncs files/folders from my phone back to my server via ssh. Some folders are on a schedule, some monitor for changes and sync immediately; most are just one-way, some are two-way (files added to the server will sync back to the phone as well as uploading data to the server). There’s even one that automatically drops files into paperless-ngx’ consume folder for automatic document importing.
From there BorgBackup makes a daily backup of the data, keeping historical backups for years with absolutely incredible efficiency. I currently have 21 backups of about ~550gb each. Borg stores this in 447gb of total disc space.
Without authentication; it’s possible to randomly generate UUIDs and use them to retrieve media from a jellyfin server. That’s about the only actually concerning issue on that list, and it’s incredibly minor IMO.
With authentication, users (ie, the people you have trusted to access your server) can potentially attack each other, by changing each others settings and viewing each other’s watch history/favorites/etc.
That’s it. These issues aren’t even worth talking about for 99.9% of jellyfin users.
Should they be fixed? Sure, eventually. But these issues aren’t cause to yell about how insecure jellyfin is in every single conversation, and to go trying to scare everyone off of hosting it publicly. Stop spreading FUD.
Can’t say I disagree.
Yeah; Emby was originally called MediaBrowser and was a free open source project. ‘MediaBrowsers’ developers decided to move to a closed source paid model to establish some more consistent income and support the dedicated developers they have. Thus Emby was born.
Some users were really unhappy with this decision and forked MediaBrowsers last release to create Jellyfin. Their development has been quite a bit slower, but they’ve made some significant strides in recent years. It’s a more and more attractive option.
One of my biggest reasons for sticking with Emby (besides already having a lifetime premier license) is the dedicated clients available on more platforms. Xbone is my primary streaming device, besides android: Emby has a dedicated xbox client you can install that will take full advantage of the the hardware(more content direct plays, HEVC video for example), where as Jellyfin you’ve gotta use the web browser which is cumbersome and forces the server to transcode media a lot more.
In the case of plex, it’s not 100% selfhosted. There’s a dependence on plexs public infrastructure for user management/authentication. They also help bypass NAT by proxying connections through their servers so you don’t have to setup port forwarding and can even easily escape double NAT situations.
I can understand paying for that convenience, but cost keeps rising while previously free features continue to get locked behind paywalls.
Tbh, having users required to authenticate with plex.tv was enough for me to look elsewhere. The biggest reason to self host for me is to remove dependency on public services.
I got the same email.
I haven’t had plex installed for over 7 years, and I’ve NEVER used the shared libraries feature.
We noticed that you’ve accessed libraries from friends and family in the past
They’ve apparently noticed activity that’s never occurred.
You could setup a user account like the share you’re describing. There’s a setting to prevent the user from changing their password.
Just pass out those credentials to anyone you want to collaborate with; they don’t need their own individual accounts.
I use https://filebrowser.org/ for this.
Nice lightweight filebrowsing/sharing with user management. Users can have their own dedicated directories, or collaborate.
You can also create share links that allow anyone with the link to view/download files. Optionally password protected.
Here’s a demo you can mess with: https://demo.filebrowser.org/ User: demo Pass: demo
Most of my web services are behind my vpn, but there are a couple I expose publicly for friends/family to use. Things like emby, ombi, and some generic file sharing with file browser.
One of these has a long custom path setup in nginx which, instead of proxying to the named service, will ask for http basic auth credentials. Use the correct host+path, then provide the correct user+pass, and you’ll be served an openvpn configuration file which includes an encrypted private key. Decrypt that and you’ve got backdoor vpn access.
I keep vaultwarden behind a vpn so it’s not exposed directly to the net. You don’t need a constant connection to the server; that’s only needed to add/change vault items.
This does require some planning though; it’s easy to lock yourself out of your accounts when you’re away, if you don’t incorporate a backdoor of some kind to let yourself in in an emergency. (lost your device while away from home for example)
My normal vpn connection requires a private key and a password that’s stored in my vault to decrypt it. I’ve setup a method for retrieving a backup set of keys using a series of usernames, emails, passwords, and undocumented paths (these are the only passwords I actually memorize); allowing me to reach vaultwarden where I can retrieve my vault with the data needed to login to everything else properly.
Usually that does the trick for me too; but this morning it just would not cooperate no matter what I tried.
Seems to be playing ball again, for now.
I have a feeling this is more to do with Android/Google not wanting to give up control more than anything. If googles stuff always works, but third party stuff is mysteriously always glitchy; users are going to gravitate to google and their ever growing monopoly…
Thank you! You gave me the hint I needed.
I didn’t know there was a quick setting button (the buttons in the notification tray) and have been struggling to find the accessibility options people have mentioned.
That button in the tray seems so much more reliable. Thanks again!
I tried. I couldn’t get it to work again, so wanted to look at other options alongside looking for help/solutions.
But just as it decided to stop working, despite my efforts; it’s suddenly started working again.
Sigh…
Vaultwarden is just a self-hosted server for Bitwardens clients. It’s Bitwardens android client I’ve been having issues with.
That’s an interesting option. It’s the Bitwarden app I’ve been having issues with; though I’m not sure how much of that is Bitwardens fault vs Android itself.
I’ll give that a look, thanks :)
An $11/yr domain pointed at my IP. Port 443 is open to nginx, which proxies to the desired service depending on subdomain. (and explicitly drops any connection that uses my raw ip or an unrecognized name to connect, without responding at all)
ACME.sh automatically refreshes my free ssl certificate every ~2months via DNS-01 verification and letsencrypt.
And finally, I’ve got a dynamic IP, so DDClient keeps my domain pointed at the correct IP when/if it changes.
There’s also pihole on the local network, replacing the WAN IP from external DNS, with the servers local IP, for LAN devices to use. But that’s very much optional, especially if your router performs NAT Hairpinning.
This setup covers all ~24 of the services/web applications I host, though most other services have some additional configuration to make them only accessible from LAN/VPN despite using the same ports and nginx service. I can go into that if there’s interest.
Only Emby/Jellyfin, Ombi, and Filebrowser are made accessible from WAN; so I can easily share those with friends/family without having to guide them through/restrict them to a vpn connection.