I’m considering a business plan for people getting in to self-hosting. Essentially I sell you a Mikrotik router and a refurbished tiny x86 server. The idea is that the router plugs in to your home internet and the server into the router. Between the two they get the server able to handle incoming requests so that you can host services on the box and address them from the broader Internet.
The hypothesis is that $150 of equipment to avoid dozens of hours of software configuration is a worthwhile trade for some customers. I realize some people want to learn particular technologies and this is a bad fit for them. I think there are people out there that want the benefit of self-hosting, and may find it worth it to buy “self-hosting in a box”.
What do you think? Would this be a useful product for some people?
Coming from someone who started selfhosting on a pi 2B (similar-ish specs), you’d be surprised. If you don’t need anything fast or fancy, that 1GB will go a long way, and plenty of selfhosted apps require very little CPU. The only real problem I faced was that all HTTPS-related network tasks were limited at ~3MB/s, as that is how fast my pi could encrypt the data (presumably, I just saw my webserver utilising the entire CPU and figured this was the most likely explanation)
I’m currently hosting like 5 vms on a proxmox host (mostly ubuntu vms- pihole, nextcloud, home assistant, etc), which is an i5 4590 with 32 gb ram and I’m running up against the limits of how much ram I can provision and if 2 or more of my vms are doing something intensive at the same time I’m pinning the CPU. I don’t think my use-case is that crazy for someone doing a little self-hosting.
Luxury! My homeserver has an i5 3470 with 6GB or RAM (yes, it’s a cursed 4+2 setup)! </badMontyPythonReference>
Interesting, I also run Nextcloud and pihole, and vaultwarden, jellyfin, paperless-ngx, gitea, vscode-server and a minecraft server (every now and then).
You’re right that such a system really does show its age, but only when doing multiple intensive tasks at the same time. I try not to backup my photos to Nextcloud while running minecraft, for example, as the imagine identification task pins my CPU at 100%. So yes, I agree, you’re probably not doing anything out of the ordinary on your setup.
The point I was trying to make still stands though, as that pi 2B could run more than I would’ve expected beforehand. I believe it once even ran jellyfin, a simple file server, samba, and a webserver with a simple HTML website. Jellyfin worked just fine, as long as the pi didn’t have to transcode (never got hardware transcoding to work).
It is funny that you should run out of memory, seeing as everything fits (albeit, just barely) on my machine in 1/5 the memory. Would de overhead of running VM’s account for such a large difference?
I’m running the recognize app on nextcloud which I think requires at least 4-5 GB RAM, so I have 6 dedicated to that VM. I’m pretty sure the recommendation for Ubuntu in general was 2 GB RAM so I gave my pihole half that. Home assistant wanted 4 GB, but I gave it 2. I think my Jellyfin server has like 6 and I have another VM with like 4. So that’s a total of like 19gb RAM provisioned. Plus I have a 2 TB zfs pool for my nextcloud VM. When I go into proxmox it tells me I’m using like 29.5 GB.
I suspect if someone was using docker or some other sort of containerization one could expect better performance than what I am getting with VMs.