this post was submitted on 03 Jan 2024
48 points (94.4% liked)

Selfhosted

37751 readers
778 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 1 year ago
MODERATORS
 

I have recently repurposed and old Hp Stream to a home server and successfully run Immich. I really like it and even a small 500GB disk is way more than the 15GB Google offers.

My issue though is about backup. I would only be comfortable if all the data is backed up in an off-site server (cloud). But the back up storage will probably cost as much as paying for a service like ente or similar, directly replacing Google photo.

What am I missing? Where do you store your backup?

top 29 comments
sorted by: hot top controversial new old
[–] butitsnotme 35 points 6 months ago (5 children)

I backup to a external hard disk that I keep in a fireproof and water resistant safe at home. Each service has its own LVM volume which I snapshot and then backup the snapshots with borg, all into one repository. The backup is triggered by a udev rule so it happens automatically when I plug the drive in; the backup script uses ntfy.sh (running locally) to let me know when it is finished so I can put the drive back in the safe. I can share the script later, if anyone is interested.

[–] [email protected] 5 points 6 months ago (1 children)

Please! That sounds like a slick setup.

[–] butitsnotme 2 points 1 month ago* (last edited 1 month ago)

I followed the guide found here, however with a few modifications.

Notably, I did not encrypt the borg repository, and heavily modified the backup script.

#!/bin/bash -ue

# The udev rule is not terribly accurate and may trigger our service before
# the kernel has finished probing partitions. Sleep for a bit to ensure
# the kernel is done.
#
# This can be avoided by using a more precise udev rule, e.g. matching
# a specific hardware path and partition.
sleep 5

#
# Script configuration
#

# The backup partition is mounted there
MOUNTPOINT=/mnt/external

# This is the location of the Borg repository
TARGET=$MOUNTPOINT/backups/backups.borg

# Archive name schema
DATE=$(date '+%Y-%m-%d-%H-%M-%S')-$(hostname)

# This is the file that will later contain UUIDs of registered backup drives
DISKS=/etc/backups/backup.disk

# Find whether the connected block device is a backup drive
for uuid in $(lsblk --noheadings --list --output uuid)
do
        if grep --quiet --fixed-strings $uuid $DISKS; then
                break
        fi
        uuid=
done

if [ ! $uuid ]; then
        echo "No backup disk found, exiting"
        exit 0
fi

echo "Disk $uuid is a backup disk"
partition_path=/dev/disk/by-uuid/$uuid
# Mount file system if not already done. This assumes that if something is already
# mounted at $MOUNTPOINT, it is the backup drive. It won't find the drive if
# it was mounted somewhere else.
(mount | grep $MOUNTPOINT) || mount $partition_path $MOUNTPOINT
drive=$(lsblk --inverse --noheadings --list --paths --output name $partition_path | head --lines 1)
echo "Drive path: $drive"

# Log Borg version
borg --version

echo "Starting backup for $DATE"

# Make sure all data is written before creating the snapshot
sync


# Options for borg create
BORG_OPTS="--stats --one-file-system --compression lz4 --checkpoint-interval 86400"

# No one can answer if Borg asks these questions, it is better to just fail quickly
# instead of hanging.
export BORG_RELOCATED_REPO_ACCESS_IS_OK=no
export BORG_UNKNOWN_UNENCRYPTED_REPO_ACCESS_IS_OK=no


#
# Create backups
#

function backup () {
  local DISK="$1"
  local LABEL="$2"
  shift 2

  local SNAPSHOT="$DISK-snapshot"
  local SNAPSHOT_DIR="/mnt/snapshot/$DISK"

  local DIRS=""
  while (( "$#" )); do
    DIRS="$DIRS $SNAPSHOT_DIR/$1"
    shift
  done

  # Make and mount the snapshot volume
  mkdir -p $SNAPSHOT_DIR
  lvcreate --size 50G --snapshot --name $SNAPSHOT /dev/data/$DISK
  mount /dev/data/$SNAPSHOT $SNAPSHOT_DIR

  # Create the backup
  borg create $BORG_OPTS $TARGET::$DATE-$DISK $DIRS


  # Check the snapshot usage before removing it
  lvs
  umount $SNAPSHOT_DIR
  lvremove --yes /dev/data/$SNAPSHOT
}

# usage: backup <lvm volume> <snapshot name> <list of folders to backup>
backup photos immich immich
# Other backups listed here

echo "Completed backup for $DATE"

# Just to be completely paranoid
sync

if [ -f /etc/backups/autoeject ]; then
        umount $MOUNTPOINT
        udisksctl power-off -b $drive
fi

# Send a notification
curl -H 'Title: Backup Complete' -d "Server backup for $DATE finished" 'http://10.30.0.1:28080/backups'

Most of my services are stored on individual LVM volumes, all mounted under /mnt, so immich is completely self-contained under /mnt/photos/immich/. The last line of my script sends a notification to my phone using ntfy.

[–] [email protected] 3 points 6 months ago (1 children)

I am super curious about the udev triggering, didn't know thats possible!

[–] butitsnotme 2 points 1 month ago

See my other reply here.

[–] [email protected] 2 points 6 months ago (1 children)

This sounds really interesting, please share.

[–] butitsnotme 1 points 1 month ago

See my other reply here.

[–] [email protected] 2 points 6 months ago (1 children)

I would love to see your script! I’m in desperate need of a better backup strategy for my video projects

[–] butitsnotme 2 points 1 month ago

See my other reply here.

[–] roofuskit 1 points 5 months ago

Fireproof safes don't protect against heat except what's high enough to combust paper. Temps will still probably be high enough to destroy a drive with a regular fireproof safe.

[–] [email protected] 10 points 6 months ago (1 children)

I use Backblaze B2 for my backups. Storing about 2tb, comes out to about $10/mo, which is on par with Google One pricing. However, I get the benefit of controlling my data, and I use it for tons more than just photos (movies/shows etc).

If you want a cheaper solution and have somewhere else you can store off-site (e.g. family/friend's house), you can probably use a raspberry pi to make a super cheap backup solution.

[–] [email protected] 9 points 6 months ago* (last edited 6 months ago) (1 children)

If you have 1tb+ of data you can get a cheaper option just by moving to hetzner (also, even storj is cheaper than backblaze)

[–] [email protected] 1 points 6 months ago

This is exactly what I was looking for last night, thank you!

[–] [email protected] 7 points 6 months ago (1 children)

I have my Immich library backed up to Backblaze B2 via Duplicacy. That job runs nightly. I also have a secondary sync to Nextcloud running on another server. That said, I need another off prem backup and will likely run a monthly job to my parents house either via manually copying to an external disk then taking it over or setting up a Pi or other low power server and a VPN to do it remotely.

[–] [email protected] 1 points 6 months ago (1 children)

So you just mount the immich folder in the duplicate container? Or run it native?

[–] [email protected] 1 points 5 months ago

Immich and Duplicacy both run on my unraid server. Duplicacy just watches the Immich pics folder and backs that up nightly.

[–] [email protected] 4 points 6 months ago

Backblaze B2 is 6$ a month for 1TB and first 10GB is free. You pay proportionally (it cost me 2-3$ for last 7-8 months for 20-150 GB that accumulated over time). Keep in mind that you will spend more if you download backup, but you should use cloud backup as last resort anyway. I backup to 2nd local disk and also to B2 daily with Kopia. Didnt need backup fortunately, downloading from B2 small files ocasionally just for testing setup

Its not just cheaper, I love it because I dont have to deal with Gshit company

[–] [email protected] 4 points 6 months ago

There was a good blog post about the real cost of storage, but I can't find it now.

The gist was that to store 1TB of data somewhat reliably, you probably need at least:

  • mirrored main storage 2TB
  • frequent/local backup space, also at least mirrored disks 2TB + more if using a versioned backup system
  • remote / cold storage backup space about the same as the frequent backups

Which amounts to something like 6TB of disk for 1TB of actual data. In real life you'd probably use some other level of RAID, at least for larger amounts so it's perhaps not as harsh, and compression can reduce the required backup space too.

I have around 130G of data in Nextcloud, and the off-site borg repo for it is about 180G. Then there's local backups on a mirrored HDD, with the ZFS snapshots that are not yet pruned that's maybe 200G of raw disk space. So 130G becomes 510G in my setup.

[–] machinin 3 points 6 months ago

I bought a Synology that I keep at my in-laws, then use Syncthing to keep my pictures backed up. I just started, so I don't know how it will go long term.

If anyone else has a better option than Syncthing for Linux to Synology, I would love to hear it.

[–] [email protected] 3 points 6 months ago* (last edited 1 month ago)

Acronyms, initialisms, abbreviations, contractions, and other phrases which expand to something larger, that I've seen in this thread:

Fewer Letters More Letters
LVM (Linux) Logical Volume Manager for filesystem mapping
RAID Redundant Array of Independent Disks for mass storage
VPN Virtual Private Network
ZFS Solaris/Linux filesystem focusing on data integrity

4 acronyms in this thread; the most compressed thread commented on today has 7 acronyms.

[Thread #401 for this sub, first seen 4th Jan 2024, 18:05] [FAQ] [Full list] [Contact] [Source code]

[–] [email protected] 2 points 6 months ago* (last edited 6 months ago) (1 children)

I use backblaze on my synology. I backup photos automatically to it with their built in app on my phone, then every night I run encryped backups. I also could setup an encrypted backup to go to my parent's synology.

My backup is about 900gb and costs <$5/mo. That is my music, pictures, movies, and TV shows. Obviously that will increase, but well worth the nominal coat to have that much backup encrypted and in the cloud.

[–] [email protected] 1 points 6 months ago (1 children)
[–] [email protected] 1 points 6 months ago

I do. And since I've been slowly taking back control over all my online stuff as much as I can, I'm very happy with it. It gives me peace of mind it's secure and I am super unlikely to just lose it.

[–] [email protected] 2 points 6 months ago

For me the answer is that I need off site backup anyway for stuff like important digital documents, passwords and more. For me a dedicated storage provider I trust far more than Google/Apple/Microsoft which all have a financial interest in understanding me and my patterns to better sell additional services too me. So I use Dropbox but if you're more technically inclined and have a lot of data then something akin to say Wasabi could make financial sense.

[–] [email protected] 1 points 5 months ago* (last edited 5 months ago)

I have a 2N+C backup strategy. I have two NASes, and I use rclone to backup my data from one NAS to the other, and then backup (with encryption) my data to Amazon S3. I have a policy on that bucket in S3 that shoves all files into Glacier Deep Archive at day 0, so I pay the cheapest rate possible.

For example, I'm storing just shy of 400GB of personal photos and videos in one particular bucket, and that's costing me about $0.77USD per month. Pennies.

Yes, it'll cost me a lot more to pull it out and, yes, it'll take a day or two to get it back. But it's an insurance policy I can rely on and a (future) price I'm willing to pay should the dire day (lost both NASes, or worse) ever arrive when I need it.

Why Amazon S3? I'm in Australia, and that means local access is important to me. We're pretty far from most other places around the world. It means I can target my nearest AWS region with my rclone jobs and there's less latency. Backblaze is a great alternative, but I'm not in the US or Europe. Admittedly, I haven't tested this theory, but I'm willing to bet that in-country speeds are still a lot quicker than any CDN that might help get me into B2.

Also, something others haven't yet mentioned is, per Immich's guidance on their repo (Disclaimer right at the top) is not NOT rely on Immich as your sole backup. Immich is under very active development, and breaking changes are a real possibility all the time right now.

So, I use SyncThing to also backup all my photos and videos to my NAS, and that's also backed up to the other NAS and S3. That's why I have nearly 400GB of photos and videos - it's effectively double my actual library size. But, again, at less than a buck a month to store all that, I don't really mind double-handling all that data, for the peace of mind I get.

[–] Telodzrum 1 points 6 months ago (1 children)

pCloud sells itself as a privacy-focused alternative to Dropbox, Google, iCloud, etc. They're running a deal right now on lifetime accounts, too.

[–] roofuskit 2 points 5 months ago (1 children)

Never trust a company selling lifetime accounts. It's entirely unsustainable and eventually the other shoe always drops.

[–] Telodzrum 2 points 5 months ago (1 children)
[–] roofuskit 1 points 5 months ago* (last edited 5 months ago)

We're talking about data storage, not software. There are real every day costs, maintenance, replacement, power, etc... that are involved in reliably storing data.

I share the sentiment that you should be able to buy software.

Paying for data storage in a single lifetime payment is like buying one square foot of storage space in someone's apartment for a flat fee and expecting it to actually be there forever.