This is the documentation of pg-backup for all old versions bellow v1.0. In the old version, S3 storage was mounted using s3fs, so we decided to migrate to the official AWS SDK.

Storage:

  • local
  • s3
  • Object storage

Volumes:

  • /s3mnt => S3 mounting path
  • /backup => local storage mounting path

Usage

Options Shorts Usage
pg-bkup bkup CLI utility
backup   Backup database operation
restore   Restore database operation
history   Show the history of backup
–storage -s Storage. local or s3 (default: local)
–file -f File name to restore
–path   S3 path without file name. eg: /custom_path
–dbname -d Database name
–port -p Database port (default: 5432)
–mode -m Execution mode. default or scheduled (default: default)
–disable-compression   Disable database backup compression
–prune   Delete old backup, default disabled
–keep-last   Delete old backup created more than specified days ago, default 7 days
–period   Crontab period for scheduled mode only. (default: “0 1 * * *”)
–help -h Print this help message and exit
–version -V Print version information and exit

Environment variables

Name Requirement Description
DB_PORT Optional, default 5432 Database port number
DB_HOST Required Database host
DB_NAME Optional if it was provided from the -d flag Database name
DB_USERNAME Required Database user name
DB_PASSWORD Required Database password
ACCESS_KEY Optional, required for S3 storage AWS S3 Access Key
SECRET_KEY Optional, required for S3 storage AWS S3 Secret Key
BUCKET_NAME Optional, required for S3 storage AWS S3 Bucket Name
S3_ENDPOINT Optional, required for S3 storage AWS S3 Endpoint
FILE_NAME Optional if it was provided from the –file flag Database file to restore (extensions: .sql, .sql.gz)

Note:

Creating a user for backup tasks who has read-only access is recommended!

create read-only user

Backup database :

Simple backup usage

bkup backup

S3

pg-bkup backup --storage s3

Docker run:

docker run --rm --network your_network_name \
--name pg-bkup -v $PWD/backup:/backup/ \
-e "DB_HOST=database_host_name" \
-e "DB_USERNAME=username" \
-e "DB_PASSWORD=password" jkaninda/pg-bkup:v0.7  pg-bkup backup -d database_name

Docker compose file:

version: '3'
services:
  postgres:
    image: postgres:14.5
    container_name: postgres
    restart: unless-stopped
    volumes:
      - ./postgres:/var/lib/postgresql/data
    environment:
      POSTGRES_DB: bkup
      POSTGRES_PASSWORD: password
      POSTGRES_USER: bkup
  pg-bkup:
    image: jkaninda/pg-bkup:v0.7
    container_name: pg-bkup
    depends_on:
      - postgres
    command:
      - /bin/sh
      - -c
      - pg-bkup backup -d bkup
    volumes:
      - ./backup:/backup
    environment:
      - DB_PORT=5432
      - DB_HOST=postgres
      - DB_NAME=bkup
      - DB_USERNAME=bkup
      - DB_PASSWORD=password

Restore database :

Simple database restore operation usage

pg-bkup restore --file database_20231217_115621.sql  --dbname database_name
pg-bkup restore -f database_20231217_115621.sql -d database_name

S3

pg-bkup restore --storage s3 --file database_20231217_115621.sql --dbname database_name

Docker run:

docker run --rm --network your_network_name \
--name pg-bkup \
-v $PWD/backup:/backup/ \
-e "DB_HOST=database_host_name" \
-e "DB_USERNAME=username" \
-e "DB_PASSWORD=password" \
jkaninda/pg-bkup:v0.7   pg-bkup restore -d database_name -f store_20231219_022941.sql.gz

Docker compose file:

version: '3'
services:
  pg-bkup:
    image: jkaninda/pg-bkup:v0.7
    container_name: pg-bkup
    command:
      - /bin/sh
      - -c
      - pg-bkup restore --file database_20231217_115621.sql -d database_name
    volumes:
      - ./backup:/backup
    environment:
      #- FILE_NAME=database_20231217_040238.sql.gz # Optional if file name is set from command
      - DB_PORT=5432
      - DB_HOST=postgres
      - DB_USERNAME=user_name
      - DB_PASSWORD=password

Run

docker-compose up -d

Backup to S3

docker run --rm --privileged \
--device /dev/fuse --name pg-bkup \
-e "DB_HOST=db_hostname" \
-e "DB_USERNAME=username" \
-e "DB_PASSWORD=password" \
-e "ACCESS_KEY=your_access_key" \
-e "SECRET_KEY=your_secret_key" \
-e "BUCKETNAME=your_bucket_name" \
-e "S3_ENDPOINT=https://s3.us-west-2.amazonaws.com" \
jkaninda/pg-bkup:v0.7  pg-bkup backup -s s3 -d database_name

To change s3 backup path add this flag : –path /my_customPath . default path is /pg-bkup

Simple S3 backup usage

pg-bkup backup --storage s3 --dbname mydatabase 
  pg-bkup:
    image: jkaninda/pg-bkup:v0.7
    container_name: pg-bkup
    privileged: true
    devices:
    - "/dev/fuse"
    command:
      - /bin/sh
      - -c
      - pg-bkup restore --storage s3 -f database_20231217_115621.sql.gz --dbname database_name
    environment:
      - DB_PORT=5432
      - DB_HOST=postgress
      - DB_USERNAME=user_name
      - DB_PASSWORD=password
      - ACCESS_KEY=${ACCESS_KEY}
      - SECRET_KEY=${SECRET_KEY}
      - BUCKET_NAME=${BUCKET_NAME}
      - S3_ENDPOINT=${S3_ENDPOINT}

Run in Scheduled mode

This tool can be run as CronJob in Kubernetes for a regular backup which makes deployment on Kubernetes easy as Kubernetes has CronJob resources. For Docker, you need to run it in scheduled mode by adding --mode scheduled flag and specify the periodical backup time by adding --period "0 1 * * *" flag.

Make an automated backup on Docker

Syntax of crontab (field description)

The syntax is:

  • 1: Minute (0-59)
  • 2: Hours (0-23)
  • 3: Day (0-31)
  • 4: Month (0-12 [12 == December])
  • 5: Day of the week(0-7 [7 or 0 == sunday])

Easy to remember format:

* * * * * command to be executed
- - - - -
| | | | |
| | | | ----- Day of week (0 - 7) (Sunday=0 or 7)
| | | ------- Month (1 - 12)
| | --------- Day of month (1 - 31)
| ----------- Hour (0 - 23)
------------- Minute (0 - 59)

At every 30th minute

*/30 * * * *

“At minute 0.” every hour

0 * * * *

“At 01:00.” every day

0 1 * * *

Example of scheduled mode

Docker run :

docker run --rm --name pg-bkup \
-v $BACKUP_DIR:/backup/ \
-e "DB_HOST=$DB_HOST" \
-e "DB_USERNAME=$DB_USERNAME" \
-e "DB_PASSWORD=$DB_PASSWORD" jkaninda/pg-bkup:v0.7  pg-bkup backup --dbname $DB_NAME --mode scheduled --period "0 1 * * *"

With Docker compose

version: "3"
services:
  pg-bkup:
    image: jkaninda/pg-bkup:v0.7
    container_name: pg-bkup
    privileged: true
    devices:
    - "/dev/fuse"
    command:
      - /bin/sh
      - -c
      - pg-bkup backup --storage s3 --path /mys3_custom_path --dbname database_name --mode scheduled --period "*/30 * * * *"
    environment:
      - DB_PORT=5432
      - DB_HOST=postgreshost
      - DB_USERNAME=userName
      - DB_PASSWORD=${DB_PASSWORD}
      - ACCESS_KEY=${ACCESS_KEY}
      - SECRET_KEY=${SECRET_KEY}
      - BUCKET_NAME=${BUCKET_NAME}
      - S3_ENDPOINT=${S3_ENDPOINT}

Kubernetes CronJob

For Kubernetes, you don’t need to run it in scheduled mode.

Simple Kubernetes CronJob usage:

apiVersion: batch/v1
kind: CronJob
metadata:
  name: bkup-job
spec:
  schedule: "0 1 * * *"
  jobTemplate:
    spec:
      template:
        spec:
          containers:
          - name: pg-bkup
            image: jkaninda/pg-bkup:v0.7
            securityContext:
              privileged: true
            command:
            - /bin/sh
            - -c
            - pg-bkup backup -s s3 --path /custom_path
            env:
              - name: DB_PORT
                value: "5432" 
              - name: DB_HOST
                value: ""
              - name: DB_NAME
                value: ""
              - name: DB_USERNAME
                value: ""
              # Please use secret!
              - name: DB_PASSWORD
                value: ""
              - name: ACCESS_KEY
                value: ""
              - name: SECRET_KEY
                value: ""
              - name: BUCKET_NAME
                value: ""
              - name: S3_ENDPOINT
                value: "https://s3.us-west-2.amazonaws.com"
          restartPolicy: Never

Authors

Jonas Kaninda