Neat
Neat
Here you go:
There could probably be some additional refactoring here, but it works for my setup. I’m using default nginx paths, so they probably look different than other installs that use custom stuff like /var/www, etc.
Use it by putting it in a shell script, make it executable, then call it:
sudo scriptName.sh 28.0.1
Replace the version with whatever version you’re upgrading to. I would highly recommend never upgrading to a .0, always wait for at least a .1 patch. I left some sleeps in the when I was debugging a while back, those are safe to remove assuming it works in your setup. I also noticed some variables weren’t quoted, I’m not a bash programmer so there’s probably some consistency issues that could be addressed if someone is OCD.
Sure! I’ll respond with a link in a bit.
As a person who used to be “the backup guy” at a company, truer words are rarely spoken. Always test the backups otherwise it’s an exercise in futility.
One of my next steps was hardening my OPNSense router as it handles all the edge network reverse proxy duties, so IDS was in the list. I’m digging into Crowdsec now, it looks like there’s an implementation for OPNsense. Thanks for the tip!
Good call. I do some backups now but I should formalize that process. Any recommendations on selfhost packages that can handle the append only functionality?
I wonder what performance impact there would be if you were to move pgsql onto bare metal with enough ram dedicated to caching all of the db data (think: i5 or i7 nuc). That’s going to be my next step with my homelab; I want to migrate everything to a single db host with a lot of RAM and M2 storage and avoid the db process replication I have going on. I have no performance complaints with NC currently, I’m running PHP cache and redis as well as image preview and imaginary.
You absolutely need to move from patch to patch and cannot just do a multiple version jump safely. You also need to validate the configs between versions, especially major release updates or you risk breaking. New features and optimizations happen and you also may need to change our update your reverse proxy configuration on update, or modify db table configuration (just puking this from memory as I’ve had to do it before). I don’t know that there’s automation for each one of those steps.
Because of that, I run nextcloud in a VM and install it from the binary package. I wrote a shell script that handles downloading, moving the files, updating permissions and copying the old config forward, symlinking and doing the upgrade. Then all I have to do is log in as administrator, check out the admin dashboard and make sure there aren’t new things I have to address in the status page. It’s a pain, but my nextcloud uses external db and redis and PHP caching so it’s not an easy out of the box setup. But it’s been solid for a long time once I adopted using this script.
This is interesting, thanks!
Exit: Sadly, they don’t support media encoding at all. So might still be useful for ML duties.
Transcoding on download seems like the easiest use case, I could use Tdarr and have one node with the GPU. But for apps like Immich that use the GPU for both transcoding (raw to jpg?) and for ML purposes (facial recognition) I’m guessing the container will have to run on the hardware where the GPU is, which means Plex and Jellyfin will also have to follow the gpu.
I’ve definitely thought about moving the GPU to a dedicated mini itx box. Wonder if I could find something rack friendly…
Using NVEnc with the current linuxserver images. The readme covers the issue.
Awesome, I’ll check it out later this evening. Thank you!
I assume tdarr will take a handoff/trigger from Radarr to operate on a file?
For conversion of videos after download
For conversion of videos after download. I don’t use tdarr. Doing what you suggest works for Sonarr, but not radar because of different base images. Two different groups maintaining those projects I guess.
Edit: this is the issue I’m speaking about in particular:
I’ll start researching what the user agents are for the various services and then work on creating a simple POC with nginx. If that actually works, I can try to put together a production quality app to handle it.
If a service was serving the webfinger, it could guess which account needed to be returned based on the requesters user agent. If the UA was mastodon, it could return the mastodon link rel, if pixelfed then return that link rel, etc.
Might be able to rig it with some more complex conditional logic and regex in nginx as a bandaid. AFAICT, the webfinger spec doesn’t really allow for this, which if true, was pretty short sighted.
I haven’t considered more in depth S2S connections. I’ll have to watch the traffic logs and see what exactly is being requested and see if all of it can be directed accordingly. I see now you commented on that issue. Also, to be clear, I’m still running the services in subdomains, but I’m trying to use user@domain.tld as the discovery account.
Looks like someone has filed this issue:
Not enough info. Those are two different things.