Vaibhav Singh

Blog - vaibhavsingh.com

Automated backups using tar on Linux

Even though I use GCP’s scheduled snapshot feature as a means to backup entire OS, for a last known good configuration for my compute instances. But it seems excessive in some cases because it doesn’t make sense to restore the entire OS for something as simple as recovering an accidentally deleted file.

Therefore, we must set a cron job which executes at regular intervals and archives the directory we care for most.

Create a shell script backup.sh in your home directory. Also create the target directory where backups will be stored.

#!/bin/bash
DATE=$(date +%Y-%m-%d-%H%M%S)
BACKUP_DIR="$HOME/backups"
SOURCE_DIR="/target/source-directory"
sudo tar -cvzpf $BACKUP_DIR/backup-$DATE.tar.gz $SOURCE_DIR

Write another bit of code to delete older archives so as not to exhaust the storage. The code below looks for files which have a modified date older than 10 from today and deletes them.

find $BACKUP_DIR -mtime +10 -type f -delete

Put all the above into a single script and give it execute permission with chmod +x backup.sh

Schedule Automated Execution

Set a cron job to execute the script at regular intervals. Install and enable cron if your environment doesn’t have it already.

sudo apt update
sudo apt install cron
sudo systemctl enable cron
crontab -e
  # Run backup.sh at 4:30am everyday
  30 4 * * * /absolute-path/backup.sh

Debugging

It’s important to test the execution in a temp directory with junk data first before attempting a first pass on your critical files. I typically set flags -xv at the shebang to see what’s going on with the script at execution stage. Eg. !/bin/bash -xv

Tags:

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top