this post was submitted on 20 Jun 2024
23 points (92.6% liked)

Selfhosted

40645 readers
277 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 2 years ago
MODERATORS
 

I'm in the process of planning upgrades for my server, and I'm trying to figure out the "best" drive configuration for Docker. My general understanding would be that the containers should be running from an SSD, and any storage (images, videos, documents) should use a volume on an HDD.

Is it as simple as changing the data-root to point to the SSD, and keep my external volumes on the HDD as defined in my existing compose files? I've already moved data-root once b/c the OverlayFS was chewing up the limited space on the OS drive, so that process isn't daunting.

If there's a better way to configure Docker, I'm open to it, as long as it doesn't require rebuilding everything from scratch.

For reference, the server is running Debian Bookworm on an older i5 3400 with 32GB RAM.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 4 points 6 months ago (2 children)

Your CPU may be a bottleneck depending on what you are doing. The slow ram speeds will mean processing will go slowly.

In my homelab I run all SSDs. They are cheap enough that I can afford them without problem. However, if you use a mix of spinning rust and SSDs you should separate them down into different pools. I would personally have the containers on SSDs and the data storage on HDDs. In the docker compose you can do a directory mount to a HDD pool.

Depending on what you are doing I would say you should get into Proxmox early. Get a small boot SSD and then create a larger SSD zfs pool and a HDD ZFS pool. From there you can setup your VMs to use either. You could have a VM with one disk in the HDD and a second disk in the SSD. This setup would also give you the flexibility to dynamically move things around. Proxmox will not work well on that old of a CPU so if you wanted to get fancy like I'm describing you would need to upgrade to something newer.

[–] [email protected] 2 points 6 months ago (1 children)

I5 3470 is old, but its not that bad. Lots of people are homelabing on NUCs which are only very slightly faster. Performance per Watt will be terrible though. (I am on an i7-10710u, and I've yet to run out of steam so far - https://cpu.userbenchmark.com/Compare/Intel-Core-i7-10710U-vs-Intel-Core-i5-3470/m900004vs2771 )

It has VTx/VTd, so should be okay for proxmox, what makes you think it won't work well?

[–] [email protected] 3 points 6 months ago (1 children)

I had in my head that it didn't have the proper extensions for virtualization.

However, the memory and core count will be a bottleneck with virtualization. Only having 4 cores will make it a hard to delicate resources and the slower ram will mean you could have performance issues. It really depends on what you are doing I suppose. It does have 6mb of cache which will help some.

If you got a i5 6500 with ddr4 memory you would have much better performance.

[–] [email protected] 1 points 6 months ago

4 cores is a bit limiting, but definitely depends on the usage. I only have 1 VM on my NUC, everything else is docker.

I thought all the core processors had VT* extensions, I was using virtualization on my first gen i7. They are very old an inefficient now though.

[–] [email protected] 1 points 5 months ago (1 children)

For the most part, this old bucket is doing just fine with probably more than I should be throwing at it.

I'm curious as to why proxmox and VMs over a minimal Debian install with docker containers, though? At least, from my understanding, proxmox would be requiring a lot more hardware overhead when I'm mostly just running emby/jellyfin, nextcloud, homeassistant (and related services) and frigate (with a coral).

It's definitely a lot, but I also rarely see cpu use over 70% (typically much lower), though frigate likes to cause problems occasionally. And I've never seen a concerning amount of ram usage.

Definitely getting one of those little n100s soon, and will probably move the home automation stuff over there, and slowly transition the current box into being a nas and nothing more.

[–] [email protected] 1 points 5 months ago (1 children)

Segmentation really. KVM doesn't have a not of overhead but with Proxmox you can separate out everything easier. Also it makes moving services between machines very easy. When moving you might end up with a few dropped packets but the VM will stay running as it moves to a different machine.

[–] [email protected] 1 points 5 months ago (1 children)

Interesting. I'll dig a little more then. Most of my vm experience has been on desktop for various reasons, and it's almost always been a pain in the ass and not worth the effort.

I assume the kvm stuff can be running a minimal os, sort of like an alpine docker image?

[–] [email protected] 1 points 5 months ago

It runs Proxmox as the base