Storage & Backups
Create Storage Pools
Although my current setup has changed over time I currently use 3 ZFS zpools in addition to Proxmox's default rpool
. Some of the best practices I try and make sure to do on all my pool…
- enable compression (it's almost universally best to use
zstd
if available) - ensure the correct
ashift
and use at least 4k blocks (ashift=12
)
Name | Disks | Descrption |
---|---|---|
zpool10 | 6x 10TB HDD | HDD pool for all mass data storage (backups, downloads, media, etc.) |
fpool | 2x 512GB SSD | Fast SSD pool for self-hosted service's persistent data (databases, configuration, etc.) |
vpool | 2x 1TB SSD | Fast SSD pool dedicated to Docker (images, volumes, etc.) |
Creation Commands
zpool10
zpool create -o ashift=12 -o failmode=continue -O compression=lz4 -O xattr=sa -O atime=off -O encryption=aes-256-gcm -O keyformat=passphrase -m /storage/zpool10 zpool10 raidz2 \
/dev/disk/by-id/ata-WDC_WD100EMAZ-00WJTA0_JEKH3DVZ \
/dev/disk/by-id/ata-WDC_WD100EMAZ-00WJTA0_2YK148SD \
/dev/disk/by-id/ata-WDC_WD100EMAZ-00WJTA0_JEKH8RWZ \
/dev/disk/by-id/ata-WDC_WD100EMAZ-00WJTA0_JEK6ESAN \
/dev/disk/by-id/ata-WDC_WD100EMAZ-00WJTA0_JEK53ZHN \
/dev/disk/by-id/ata-WDC_WD100EMAZ-00WJTA0_2YK0HL0D
ashift=12
to use 4k blocksfailmode=continue
to let us keep reading if a drive goes badcompression=lz4
save save space and increase speedxattr=sa
be more efficient for linux attributesatime=off
don't waste time tracking access times for filesencryption=aes-256-gcm
use fast/secure encryption algorithmkeyformat=passphrase
unlock with a passphrase
fpool
zpool create -o ashift=12 -O compression=zstd -O xattr=sa -O atime=off -O encryption=aes-256-gcm -O keyformat=passphrase -m /storage/fpool fpool mirror \
/dev/disk/by-id/nvme-SAMSUNG_MZVLB512HAJQ-000H2_S498NE0K905004 \
/dev/disk/by-id/nvme-SAMSUNG_MZVLB512HAJQ-000H2_S498NE0K905006
ashift=12
to use 4k blockscompression=zstd
save save space and increase speedxattr=sa
be more efficient for linux attributesatime=off
don't waste time tracking access times for filesencryption=aes-256-gcm
use fast/secure encryption algorithmkeyformat=passphrase
unlock with a passphrase
vpool
zpool create -o ashift=12 -O compression=zstd -O xattr=sa -O atime=off -O encryption=aes-256-gcm -O keyformat=passphrase -m /storage/vpool vpool mirror \
/dev/disk/by-id/nvme-PCIe_SSD_19081510242795 \
/dev/disk/by-id/nvme-PCIe_SSD_19081510243764
ashift=12
to use 4k blockscompression=zstd
save save space and increase speedxattr=sa
be more efficient for linux attributesatime=off
don't waste time tracking access times for filesencryption=aes-256-gcm
use fast/secure encryption algorithmkeyformat=passphrase
unlock with a passphrase
Setup ZFS Scrub (Data Integrity)
Automate ZFS scrubbing so the data integrity on disks is actively monitored, repaired if necessary, and I'm alerted if there is a problem with my disks.
Create systemd Service/Timer (source)
Create a simple systemd service template for scrubbing ZFS pools.
# /etc/systemd/system/zpool-scrub@.service
+ [Unit]
+ Description=Scrub ZFS Pool
+ Requires=zfs.target
+ After=zfs.target
+
+ [Service]
+ Type=oneshot
+ ExecStartPre=-/usr/sbin/zpool scrub -s %I
+ ExecStart=/usr/sbin/zpool scrub %I
Then create a systemd timer template for periodically running that service. I am running the scrub weekly, but semi-monthly or monthly would almost certainly be ok too.
# /etc/systemd/system/zpool-scrub@.timer
+ [Unit]
+ Description=Scrub ZFS pool weekly
+
+ [Timer]
+ OnCalendar=weekly
+ Persistent=true
+
+ [Install]
+ WantedBy=timers.target
Enable ZFS Scrub
systemctl daemon-reload
systemctl enable --now zpool-scrub@rpool.timer
systemctl enable --now zpool-scrub@zpool10.timer
systemctl enable --now zpool-scrub@fpool.timer
systemctl enable --now zpool-scrub@vpool.timer
Setup Storage Layout
I wanted to logically device my storage pool up into datasets based on their intended usage. This allows me to tweak their paramters if needed and have different snapshot/backup policies.
zpool10/backups
place to store disk and time machine backupszpool10/downloads
storage for downloadszpool10/downloads/incomplete
landing zone for incomplete downloads (recordsize=1M
for bittorrent)zpool10/media
storage for audio/tv/movieszpool10/proxmox
additional storage for proxmoxzpool10/proxmox/backups
backup for proxmox instances (in subdirectories by hostname)zpool10/services
storage for services (possibly databases, so userecordsize=16k
)
zfs create zpool10/backups
zfs create zpool10/downloads
zfs create -o recordsize=16K zpool10/downloads/incomplete
zfs create zpool10/media
zfs create zpool10/proxmox
zfs create zpool10/proxmox/backups
zfs create -o recordsize=16K zpool10/services
Setup Sanoid/Syncoid (Data Backup)
Run Sanoid for automating snapshots and Syncoid for remote backups. Unfortunately this isn't available in repositories so you have to build it yourself. However the author makes it fairly simple.
Install (source)
apt-get install debhelper libcapture-tiny-perl libconfig-inifiles-perl pv lzop mbuffer
sudo git clone https://github.com/jimsalterjrs/sanoid.git
cd sanoid
ln -s packages/debian .
dpkg-buildpackage -uc -us
apt install ../sanoid_*_all.deb
Configure Sanoid
I want to take hourly snapshots of both of my ZFS pools because sometimes I am not as careful or thoughtful as I should be about what I am doing at any given moment. But I don't want to snapshot zpool/backups
because it is a backup destination that will likely already have snapshots (rpool
snapshots are stored there for example) and I also don't want to snapshot zpool/downloads
because there is nothing important under there and it is likely to change frquently.
# /etc/sanoid/sanoid.conf
+ [template_proxmox]
+ frequently = 0
+ hourly = 24
+ daily = 7
+ weekly = 4
+ monthly = 1
+ yearly = 0
+ autosnap = yes
+ autoprune = yes
+
+ [rpool]
+ use_template = template_proxmox
+ process_children_only = yes
+ recursive = yes
+
+ [rpool/ROOT]
+ use_template = rpool
+ process_children_only = yes
+ recursive = yes
+
+ [rpool/data]
+ use_template = template_proxmox
+ weekly = 1
+ monthly = 1
+ process_children_only = yes
+ recursive = yes
Maybe this is a sin, but I'd like my snapshots to be in local time so I don't have to do the (admittedly simple) conversion in my head.
# /usr/lib/systemd/system/sanoid.service
[Service]
- Environment=TZ=UTC
+ Environment=TZ=EST
Eanble Sanoid
systemctl daemon-reload
systemctl enable --now sanoid.service
Configure Syncoid
Backup rpool
to zpool10/proxmox/backups/blackbox
Right now rpool
is just running on a single non-redundant 512GB SSD disk. Even though it is only used for Proxmox (config, templates, ISOs) this isn't great practice and I'll work on this in the future. But in the meantime I am backing up everything on a daily timer to my main ZFS pool so I could recover very quickly if the SSD dies.
# /etc/systemd/system/rpool-backup.timer
+ [Unit]
+ Description=Backup rpool daily
+
+ [Timer]
+ OnCalendar=daily
+ Persistent=true
+
+ [Install]
+ WantedBy=timers.target
# /etc/systemd/system/rpool-backup.service
+ [Unit]
+ Description=Use syncoid to backup rpool to zpool10/proxmox/backups/blackbox
+ Requires=zfs.target
+ After=zfs.target
+
+ [Service]
+ Type=oneshot
+ ExecStart=/usr/sbin/syncoid --force-delete --recursive rpool zpool10/proxmox/backups/blackbox/rpool
Backup zpool10/services
offsite
All my docker containers store their configuration and data under the zpool10/services
dataset. It is imperative this is backed up offsite so if anything catastrophic ever happens I don't lose anything important and can get back up and running as quickly as I can download my backup.
# /etc/systemd/system/zpool10-services-backup.timer
+ [Unit]
+ Description=Backup zpool10/services daily
+
+ [Timer]
+ OnCalendar=daily
+ Persistent=true
+
+ [Install]
+ WantedBy=timers.target
# /etc/systemd/system/zpool10-services-backup.service
+ [Unit]
+ Description=Use syncoid to backup zpool10/services to backedup.swigg.net:bpool/zpool10/services
+ Requires=zfs.target
+ After=zfs.target
+
+ [Service]
+ Type=oneshot
+ ExecStart=/usr/sbin/syncoid --force-delete --recursive zpool10/services root@backedup.swigg.net:bpool/zpool10/services
Enable Syncoid
systemctl daemon-reload
systemctl enable --now rpool-backup.timer
Setup Restic to Backblaze B2 (Data Backup)
Read more about setting up Restic at https://fedoramagazine.org/automate-backups-with-restic-and-systemd/
Setup Restic Backup
Create the .service
and .timer
for the backup service so that it runs everyday.
# /etc/systemd/system/restic-backup.service
+ [Unit]
+ Description=Restic backup service
+
+ [Service]
+ Type=oneshot
+ ExecStart=restic backup --verbose --tag systemd.timer $BACKUP_EXCLUDES $BACKUP_INCLUDES $BACKUP_PATHS
+ ExecStartPost=restic forget --verbose --tag systemd.timer --group-by "paths,tags" --keep-daily $RETENTION_DAYS --keep-weekly $RETENTION_WEEKS --keep-monthly $RETENTION_MONTHS --keep-yearly $RETENTION_YEARS
+ EnvironmentFile=/etc/systemd/system/restic-backup.service.d/restic-backup.conf
# /etc/systemd/system/restic-backup.timer
+ [Unit]
+ Description=Backup with restic daily
+
+ [Timer]
+ OnCalendar=daily
+ Persistent=true
+
+ [Install]
+ WantedBy=timers.target
# /etc/systemd/system/restic-backup.service.d/restic-backup.conf
+ BACKUP_PATHS="/storage/zpool10"
+ BACKUP_EXCLUDES="--exclude-file /etc/systemd/system/restic-backup.service.d/restic-excludes.txt --exclude-if-present .exclude_from_backup"
+ BACKUP_INCLUDES="--files-from /etc/systemd/system/restic-backup.service.d/restic-includes.txt"
+ RETENTION_DAYS=7
+ RETENTION_WEEKS=4
+ RETENTION_MONTHS=6
+ RETENTION_YEARS=3
+ B2_ACCOUNT_ID=xxx
+ B2_ACCOUNT_KEY=xxx
+ RESTIC_REPOSITORY=b2:swigg-backup-blackbox:storage/zpool10
+ RESTIC_PASSWORD=xxx
Setup Restic Prune
The backup command above forgets about files when they expire, but to actually delete them once they aren't referenced anymore you need to prune them. The following creates a .service
and .timer
for a prune job to be run every month.
# /etc/systemd/system/restic-prune.service
+ [Unit]
+ Description=Restic backup service (data pruning)
+
+ [Service]
+ Type=oneshot
+ ExecStart=restic prune
+ EnvironmentFile=/etc/systemd/system/restic-backup.service.d/restic-backup.conf
# /etc/systemd/system/restic-prune.timer
+ [Unit]
+ Description=Prune data from the restic repository monthly
+
+ [Timer]
+ OnCalendar=monthly
+ Persistent=true
+
+ [Install]
+ WantedBy=timers.target