

Kodi is a graphical app, like Firefox, so you won’t use docker for it.
Kodi is a graphical app, like Firefox, so you won’t use docker for it.
I have Jellyfin running for years too and it has never broken for me, I use Linuxserver image, so maybe they delay the updates a bit?.. Now, Immich has broken so many times that nowadays is the only docker I don’t keep at latest (and I know using latest is a bad practice, I understand the reasons, but the convenience of not worrying about the versions beats all that for me)
Configuration is much easier, e.g. this is the full config you need to expose nextcloud on nextcloud.example.com
(assuming caddy can reach nextcloud using the hostname nextcloud
)
nextcloud.example.com {
reverse_proxy nextcloud
}
Comparing that to ngnix configs that need a template for each different service (although to be fair they’re mostly the same).
Yeah, it’s probably a legal thing, rreading-glasses is just metadata for books, completely legal, but readarr legality is less clear, so maybe they’re trying to prevent issues.
Also I didn’t understand what is rreading-glasses and why you need it
Say you want to grab a book by Isaac Asimov, you type the name of the book in readarr search bar, readarr contacts a metadata provider to show you cover images, author, date, etc. Then when you select the book readarr uses that metadata to search for downloads and ensure you’re getting the correct book and not another random book with the same name.
The problem is that readarr uses a closed source API for it’s metadata, and it’s constantly offline, which makes it impossible to use readarr. Luckily they allow you to customize the URL for the API, and rreading-glasses is an open source implementation of that API that you can use as a drop in replacement.
Yup, but most of that is easily solvable by being consistent, e.g. always use lowercase and your email (even if it’s not the login for that site). But yes, you need to know to be consistent so it’s a good point to make.
I noticed that my Ansible playbook failed to do a docker pull on readarr, I just commented it and was going to investigate further today. This sucks, especially because rreading-glasses did in fact completely solve the issue they’re facing. Not sure why they didn’t consider migrating to it officially, it’s only a config change.
It’s strange how I never see this mentioned anywhere, but there’s a way to get unique secure passwords for every site/app without needing to store them anywhere. It’s called LessPass, and essentially generates passwords based on 3 fields (site, username, master password) and works relatively well, because the advantages are quite obvious I’ll list the potential downsides:
I used to do that, but eventually found out about https://silverbullet.md/ it’s still just markdown files but I can edit them on the go with my phone.
Plus it has some nice extra syntax to query documents that’s quite handy.
Giving it write access to a folder is not even remotely on the same level as giving it control over the host networking. Worst case scenario in the volume access is to delete that data, which is on a btrfs drive and has backups, worst case scenario for network host is root access to host machine.
Yes I am, but I don’t want to give full control of my network drive to a closed source application because it paywalled me out of being able to access my media on my local network. It’s ridiculous that I have to do that. It breaks ECI, and is a security risk. And yeah, it’s a bit paranoid, but the fact that they can fix it with a simple config and put that behind a paywall is VERY worrisome, so I now need to pay if I want to isolate Plex from the host where it’s running.
Yup, that’s exactly the problem I have, it’s ridiculous that it doesn’t let me stream from a local network just because it thinks that it’s local network is only the docker one, it should be fairly simple for Plex to figure out it’s accessible via a direct connection and it doesn’t need to route through the Plex servers for this. But it won’t get fixed because it pushes people to pay, hell from what people are saying here the config to fix this is paywalled so they create a problem for which they sell you the solutions.
Therefore it’s literally impossible for me to watch my media locally, way to go Plex.
I don’t have that configuration:
I can’t speak for OP, but I self host lots of stuff, have literally dozens of services running, have an Ansible repo to manage it all and routi some stuff through a VPS, not to mention my day job has included managing services in one way or another for a long while. This is to say, I know what I’m doing. I couldn’t setup Plex to work the way I wanted to, they expect it to run in a docker with network set to host mode, I couldn’t find any way to tell Plex that my living room TV was in the same network, it just wouldn’t accept any connections as local. I know I shot myself in the foot here by not letting it run with network on host mode, but I shouldn’t have to, the port was exposed, I could reach it through the local network IP, but I wasn’t able to stream any content locally.
I’ve had that happen to me with plex, it was probably 100% my fault because I specifically changed things during the setup of the docker file, but apparently Plex can’t figure out that is local if it’s running inside docker with non-host network, it probably only accepts local connections from the docker network, and I was never able to make it treat my actual home network as local.
Not really, if you use Kodi the information on what you have watched remains on the PC running Kodi, if you always watch from the same device that’s not a big deal, but if you like to watch stuff on your smart tv, then on your PC, and downloading some to watch on your phone on the go, having the information of which episodes you’ve watched on the server helps keep things organized.
Lots of questions, let’s take it one step at a time. You have a domain, now you can point it to your public IP, so that whenever someone tries to access example.com they ask their DNS server and it replies with 10.172.172.172 (which btw is not a valid public IP). Now that request will hit your router, you need to configure your router to redirect ports 80 and 443 to 192.168.200.101, that way the request to example.com gets to your local machine.
Ok, so now you need your local machine to reply on that port, I recommend using Caddy it’s very easy to setup, but NGIX is the more traditional approach. A simple Caddy config would look like:
example.com {
respond "Hello"
}
jellyfin.example.com {
handle {
reverse_proxy http://192.168.200.101:1020/
}
}
So after the request reaches Caddy it will see that the person tried to access, example.com and respond with a “Hello”.
If instead you had tried jellyfin.example.com the DNS would have sent you to 10.172.172.172, your router would send that to 192.168.200.101, Caddy would then send it to 192.168.200.101:1020, which is Jellyfin so that would get returned.
There are some improvements that can be made, for example if both caddy and Jellyfin are docker you can share a network between them so Jellyfin is only exposed through caddy. Another possibly good idea is to add some authentication service like Authelia or Authentik to harden stuff a little bit. Also as you might have noticed Caddy can forward stuff to other computers, so you can have one machine on your network exposing multiple services on multiple machines.
What problem are you having? Docker is very straightforward, just copy the compose file and run a command.