this post was submitted on 04 Jul 2023
68 points (98.6% liked)

Selfhosted

40407 readers
239 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 2 years ago
MODERATORS
 

I have a home server that I’m using and hosting files on it. I’m worried about it breaking and loosing access to the files. So what method do you use to backup everything?

top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 35 points 1 year ago* (last edited 1 year ago) (4 children)

Backblaze on a B2 account. 0.005$ per gb. You pay for the storage you use. You pay for when you need to download your backup.

On my truenas server, it's easy as pie to setup and easy as 🥧 to restore a backup when needed.

[–] [email protected] 4 points 1 year ago* (last edited 1 year ago)

I'll add to this that restic works amazingly with Backblaze. Plus a dozen or so other backup options.

[–] [email protected] 3 points 1 year ago

B2 is awesome. I have Duplicati set up on OpenMediaVault to backup my OS nightly to B2 (as well as a local copy to the HDD).

[–] burndown 2 points 1 year ago (4 children)

Maybe I'm stupid, but what is B2? A Backblaze product?

load more comments (4 replies)
[–] [email protected] 2 points 1 year ago

I also recommend B2, it’s an S3 compatible service so any backup software/scripts/plugins that work with S3 should work with Backblaze.

[–] [email protected] 11 points 1 year ago
[–] [email protected] 11 points 1 year ago* (last edited 1 year ago) (1 children)

You guys back up your server?

load more comments (1 replies)
[–] [email protected] 8 points 1 year ago
[–] [email protected] 8 points 1 year ago

cronjobs with rsync to a Synology NAS and then to Synology's cloud backup.

[–] [email protected] 8 points 1 year ago (1 children)

Autorestic, nice wrapper for restic.

Data goes from one server to second server, and vice versa (different provider, different geolocation). And to backblaze B2 - as far as I know cheapest s3-like storage

[–] [email protected] 4 points 1 year ago

Wasabi might also be worth mentioning, a while back I compared S3-compatible storage providers and found them to be cheaper for volumes >1TB. They now seem to be slightly more expensive (5.99$ vs. 5$), but they don't charge for download traffic.

[–] TheWozardOfIz 6 points 1 year ago (4 children)
load more comments (4 replies)
[–] humancrayon 5 points 1 year ago* (last edited 1 year ago)

I have everything in its own VM, and Proxmox has a pretty awesome built in backup feature. Three different backups (one night is to my NAS, next night to an on-site external, next night to an external that's swapped out with one at work - weekly). I don't backup the Proxmox host because reinstalling it should it die completely is not a big deal. The VM's are the important part.

I have a mini PC I use to spot check VM backups once a month (full restore on its own network, check its working, delete the VM after).

My Plex NAS only backs up the movies I really care about (everything else I can "re-rip from my DVD collection").

[–] [email protected] 5 points 1 year ago

Restic to multiple repositories, local and remote.

[–] [email protected] 5 points 1 year ago

Various different ways for various different types of files.

Anything important is shared between my desktop PC's, servers and my phone through Syncthing. Those syncthing folders are all also shared with two separate servers (in two separate locations) with hourly, daily, weekly, monthly volume snapshotting. Think your financial administration, work files, anything you produce, write, your main music collection, etc... It's also a great way to keep your music in sync between your desktop PC and your phone.

Servers have their configuration files, /etc, /var/log, /root, etc... rsynced every 15 minutes to the same two backup servers, also to snapshotted volumes. That way, should any one server burn down, I can rebuild it in a trivial amount of time. This also goes for user profiles, document directories, ProgramData, and anything non-synced on windows PC's.

Specific data sets, like database backups, repositories and such are also generally rsynced regularly, some to snapshotted volumes, some to regulars, depending on the size and volatility of the data.

Bigger file shares, like movies, tv-shows, etc... I don't backup, but they're stored on a distributed GlusterFS, so if any one server goes down, that doesn't lose me everything just yet.

Hardware will fail, sooner or later. You should see any one device as essentially disposable, and have anything of worth synced and archived automatically.

[–] [email protected] 5 points 1 year ago (1 children)

ITT: lots of the usual paranoid overkill. If you do rsync with the --backup switch to a remote box or a VPS, that will cover all bases in the real world. The probability of losing anything is close to 0.

The more serious risk is discovering that something broke 3 weeks ago and the backups were not happening. So you need to make sure you are getting some kind of notification when the script completes successfully.

[–] [email protected] 2 points 1 year ago

While I don't agree that using something like restic is overkill you are very right that backup proess monitoring is very overlooked. And recovering with the backup system of your choice is too.

I let my jenkins run the backup jobs as I have it running anyways for development tasks. When a job fails it notifies me immediately via email and I can also manually check in the web ui how the backup went.

[–] [email protected] 4 points 1 year ago

Borgbackup to cloud storage on rsync.net

[–] [email protected] 4 points 1 year ago (1 children)

3-2-1

Three copies. The data on your server.

  1. Buy a giant external drive and back up to that.

  2. Off site. Backblaze is very nice

How to get your data around? Free file sync is nice.

Veeeam community version may help you too

[–] [email protected] 3 points 1 year ago (1 children)

I'm not sure how you understand the 3-2-1 rule given how you explained it, even though you're stating the right stuff (I'm confused about your numbered list..) so just for reference for people reading that, it means that your backups need to be on:

  • 3 copies
  • 2 mediums
  • 1 offsite location
load more comments (1 replies)
[–] [email protected] 3 points 1 year ago (2 children)

Rsnapshot on a second server, saving 7 daily backups, 4 weekly backups, and 6 mk they backups

load more comments (2 replies)
[–] [email protected] 3 points 1 year ago (3 children)

My server runs Plex and has almost 50 TB of video on it. After looking at all the commercial backup options I gave up on backing up that part of the data. :-(

I do backup my personal data, which is less than a terrabyte at this point. I worked out an arrangement with a friend who also runs a server. We each have a drive in the other's server that we use for backup. Every night cron runs a simple rsync script to do an incremental backup of everything new to the other machine.

This approach cost nothing beyond getting the drives. And we will still have our data even if one of the servers is physically destroyed and unrecoverable.

[–] [email protected] 2 points 1 year ago

I also have a decent amount of video data for Plex (not nearly 50TB, but more than I want I pay to backup). I figure if worst comes to worst I can rip DVD/BluRays again (though I’d rather not) so I only backup file storage from my NAS that my laptops and desktop backup to. It’s just not worth the cost to backup data that’s fairly easy to replace.

load more comments (2 replies)
[–] [email protected] 3 points 1 year ago

Borgbase to borgbase

Rock solid for years

[–] [email protected] 3 points 1 year ago* (last edited 1 year ago)

Borgbackup, using borgmatic as a frontend, to a storage VPS. I backup dozens of machines this way. I simply add a user account for each machine on the VPS, then each machine backs up over ssh to its own account.

[–] [email protected] 3 points 1 year ago

I use proxmox server and proxmox backup server (in a VM 🫣) to do encrypted backups.

A raspberry pi has ssh access to PBS and it rsync all the files, and then uploads them to backblaze using rclone.

https://2.5admins.com/ recommended "pull" backups, so if someone hacks your server they don't have access to your backups. If the pi is hacked it can mess with everything, but the idea is that is has a smaller attack surface (just ssh).

PS. If you rclone a lot of files to backblaze use https://rclone.org/docs/#fast-list , or else it will get expensive

[–] [email protected] 3 points 1 year ago* (last edited 1 year ago)

For config files, I use tarsnap. Each server has its own private key, and a /etc/tarsnap.list file which list the files/directories to backup on it. Then a cronjob runs every week to run tarsnap on them. It's very simple to backup and restore, as your backups are simply tar archives. The only caveat is that you cannot "browse" them without restoring them somewhere, but for config files it's pretty quick and cheap.

For actual data, I use a combination of rclone and dedup (because I was involved in the project at some point, but it's similar to Borg). I sync it to backblaze because that's the cheapest storage I could find. I use dedup to encrypt the backup before sending it to backblaze though. Restoration is very similar to tarsnap:

dup-unpack -k keyfile snapshot-yyyymmdd | tar -C / -x [files..] .

Most importantly, I keep a note on how to backup/restore: Backup 101

[–] [email protected] 3 points 1 year ago

Hourly backups with Borg, nightly syncs to B2. I've been playing around with zfs snapshots also, but I don't rely on them yet

[–] [email protected] 3 points 1 year ago

restic backup to Azure and Backblaze

[–] [email protected] 2 points 1 year ago

I use Bacula to an external drive, it was a pain in the ass to configure but once it's running its super reliable and easily extended to other drives or folders

[–] [email protected] 2 points 1 year ago

btrfs send/receive to my NAS.

[–] [email protected] 2 points 1 year ago

Proxmox backs up the VMs -> backups are uploaded to the cloud.

[–] [email protected] 2 points 1 year ago

Using ESXi as a hypervisor , so I rely on Veeam. I have copy jobs to take it from local to an external + a copy up to the cloud.

[–] [email protected] 2 points 1 year ago
[–] [email protected] 2 points 1 year ago* (last edited 1 year ago)

My server is a DiskStation, so I use HyperBackup to do an encrypted backup of the important data to their Synology C2 service every night.

[–] [email protected] 2 points 1 year ago

If you are using kubernetes, you can use longhorn to provision PVCs. It offers easy S3 backup along with snapshots. It has saved me a few times.

[–] [email protected] 2 points 1 year ago

I run everything in containers, so I rsync my entire docker directory to my NAS, which in turn backs it up to the cloud.

[–] [email protected] 2 points 1 year ago* (last edited 1 year ago)

I use duplicacy to backup to my local NAS and to Storj.io. In case of a fire I'm always able to restore my files. Storj.io is cheap, easy to access from any location and your files are stored and duplicated on multiple different locations.

I have used duplicity before but restoring from a new installation takes a while, as duplicity has to reanalyze the storage.

[–] [email protected] 2 points 1 year ago

Running a Duplicacy container backing up to Google drive for some stuff and Backblaze for mostly all other data. Been using it for a couple years with no issues. The GUI and scheduling is really nice too.

[–] [email protected] 2 points 1 year ago (1 children)

A simple script using duplicity to FTP data on my private website with infinite storage. I can't say if it's good or not. It's my first time doing it.

[–] [email protected] 1 points 1 year ago

How do you have infinite storage? Gsuite?

[–] [email protected] 2 points 1 year ago

All my backups are in /home/Ryan/Documents. Please don't break my Minecraft server.

[–] [email protected] 2 points 1 year ago
  • kopia backup to 2nd disk
  • kopia backup to B2 cloud
  • duplicaty backup to google drive (only most important folder <1GB)

Most of the files are actually nextcloud so I get one more copy of files (not backup) on PC by syncing with nextcloud app

[–] [email protected] 1 points 1 year ago

Bash scripting and rclone personally, here is a video that helps https://youtu.be/wUXSLmGAtgQ

[–] [email protected] 1 points 1 year ago

Compressed pg_dump rsync’ed to off-site server.

[–] [email protected] 1 points 1 year ago

Veeam backup and recovery notnfor retail license covers up to 10 workloads. I then s3 offsite to backblaze

[–] [email protected] 1 points 1 year ago

Zfs z2 pool . Not a perfect backup, but it covers disk failure (already lost one disk with no data loss), and accidental file deletion. I'm vulnerable to my house burning down, but overall I sleep well enough.

load more comments
view more: next ›