barman-cloud-backup - Backup a PostgreSQL instance and stores it in the Cloud
barman-cloud-backup [
OPTIONS]
DESTINATION_URL SERVER_NAME
This script can be used to perform a backup of a local PostgreSQL instance and
ship the resulting tarball(s) to the Cloud. Currently AWS S3, Azure Blob
Storage and Google Cloud Storage are supported.
It requires read access to PGDATA and tablespaces (normally run as postgres
user). It can also be used as a hook script on a barman server, in which case
it requires read access to the directory where barman backups are stored.
If the arguments prefixed with --snapshot- are used, and snapshots are supported
for the selected cloud provider, then the backup will be performed using
snapshots of the disks specified using --snapshot-disk arguments. The backup
label and backup metadata will be uploaded to the cloud object store.
This script and Barman are administration tools for disaster recovery of
PostgreSQL servers written in Python and maintained by EnterpriseDB.
IMPORTANT: the Cloud upload process may fail if any file with a size
greater than the configured --max-archive-size is present either in the data
directory or in any tablespaces. However, PostgreSQL creates files with a
maximum size of 1GB, and that size is always allowed, regardless of the
max-archive-size parameter.
-
usage: barman-cloud-backup [-V] [--help] [-v | -q] [-t]
[--cloud-provider {aws-s3,azure-blob-storage,google-cloud-storage}]
[--endpoint-url ENDPOINT_URL] [-P PROFILE]
[--read-timeout READ_TIMEOUT]
[--credential {azure-cli,managed-identity}]
[-z | -j | --snappy] [-h HOST] [-p PORT] [-U USER]
[--immediate-checkpoint] [-J JOBS]
[-S MAX_ARCHIVE_SIZE] [-d DBNAME] [-n BACKUP_NAME]
[--snapshot-instance SNAPSHOT_INSTANCE]
[--snapshot-disk NAME]
[--snapshot-zone SNAPSHOT_ZONE]
[--snapshot-gcp-project SNAPSHOT_GCP_PROJECT]
[--tags [TAGS [TAGS ...]]] [-e {AES256,aws:kms}]
[--encryption-scope ENCRYPTION_SCOPE]
destination_url server_name
This script can be used to perform a backup of a local PostgreSQL instance and
ship the resulting tarball(s) to the Cloud. Currently AWS S3, Azure Blob
Storage and Google Cloud Storage are supported.
positional arguments:
destination_url URL of the cloud destination, such as a bucket in AWS
S3. For example: `s3://bucket/path/to/folder`.
server_name the name of the server as configured in Barman.
optional arguments:
-V, --version show program's version number and exit
--help show this help message and exit
-v, --verbose increase output verbosity (e.g., -vv is more than -v)
-q, --quiet decrease output verbosity (e.g., -qq is less than -q)
-t, --test Test cloud connectivity and exit
--cloud-provider {aws-s3,azure-blob-storage,google-cloud-storage}
The cloud provider to use as a storage backend
-z, --gzip gzip-compress the WAL while uploading to the cloud
-j, --bzip2 bzip2-compress the WAL while uploading to the cloud
--snappy snappy-compress the WAL while uploading to the cloud
-h HOST, --host HOST host or Unix socket for PostgreSQL connection
(default: libpq settings)
-p PORT, --port PORT port for PostgreSQL connection (default: libpq
settings)
-U USER, --user USER user name for PostgreSQL connection (default: libpq
settings)
--immediate-checkpoint
forces the initial checkpoint to be done as quickly as
possible
-J JOBS, --jobs JOBS number of subprocesses to upload data to cloud storage
(default: 2)
-S MAX_ARCHIVE_SIZE, --max-archive-size MAX_ARCHIVE_SIZE
maximum size of an archive when uploading to cloud
storage (default: 100GB)
-d DBNAME, --dbname DBNAME
Database name or conninfo string for Postgres
connection (default: postgres)
-n BACKUP_NAME, --name BACKUP_NAME
a name which can be used to reference this backup in
commands such as barman-cloud-restore and barman-
cloud-backup-delete
--snapshot-instance SNAPSHOT_INSTANCE
Instance where the disks to be backed up as snapshots
are attached
--snapshot-disk NAME Name of a disk from which snapshots should be taken
--snapshot-zone SNAPSHOT_ZONE
Zone of the disks from which snapshots should be taken
--tags [TAGS [TAGS ...]]
Tags to be added to all uploaded files in cloud
storage
Extra options for the aws-s3 cloud provider:
--endpoint-url ENDPOINT_URL
Override default S3 endpoint URL with the given one
-P PROFILE, --profile PROFILE
profile name (e.g. INI section in AWS credentials
file)
--read-timeout READ_TIMEOUT
the time in seconds until a timeout is raised when
waiting to read from a connection (defaults to 60
seconds)
-e {AES256,aws:kms}, --encryption {AES256,aws:kms}
The encryption algorithm used when storing the
uploaded data in S3. Allowed values:
'AES256'|'aws:kms'.
Extra options for the azure-blob-storage cloud provider:
--credential {azure-cli,managed-identity}
Optionally specify the type of credential to use when
authenticating with Azure Blob Storage. If omitted
then the credential will be obtained from the
environment. If no credentials can be found in the
environment then the default Azure authentication flow
will be used
--encryption-scope ENCRYPTION_SCOPE
The name of an encryption scope defined in the Azure
Blob Storage service which is to be used to encrypt
the data in Azure
Extra options for google-cloud-storage cloud provider:
--snapshot-gcp-project SNAPSHOT_GCP_PROJECT
GCP project under which disk snapshots should be stored
For Boto:
- •
- https://boto3.amazonaws.com/v1/documentation/api/latest/guide/configuration.html
For AWS:
- •
- https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-getting-set-up.html
- •
- https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-getting-started.html.
For Azure Blob Storage:
- •
- https://docs.microsoft.com/en-us/azure/storage/blobs/authorize-data-operations-cli#set-environment-variables-for-authorization-parameters
- •
- https://docs.microsoft.com/en-us/python/api/azure-storage-blob/?view=azure-python
For libpq settings information:
- •
- https://www.postgresql.org/docs/current/libpq-envars.html
For Google Cloud Storage: * Credentials:
https://cloud.google.com/docs/authentication/getting-started#setting_the_environment_variable
Only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported at the
moment.
If using --cloud-provider=aws-s3:
- •
- boto3
If using --cloud-provider=azure-blob-storage:
- •
- azure-storage-blob
- •
- azure-identity (optional, if you wish to use
DefaultAzureCredential)
If using --cloud-provider=google-cloud-storage * google-cloud-storage
If using --cloud-provider=google-cloud-storage with snapshot backups
- •
- grpcio
- •
- google-cloud-compute
- 0
- Success
- 1
- The backup was not successful
- 2
- The connection to the cloud provider failed
- 3
- There was an error in the command input
- Other non-zero codes
- Failure
This script can be used in conjunction with post_backup_script or
post_backup_retry_script to relay barman backups to cloud storage as follows:
-
post_backup_retry_script = 'barman-cloud-backup [*OPTIONS*] *DESTINATION_URL* ${BARMAN_SERVER}'
When running as a hook script, barman-cloud-backup will read the location of the
backup directory and the backup ID from BACKUP_DIR and BACKUP_ID environment
variables set by barman.
Barman has been extensively tested, and is currently being used in several
production environments. However, we cannot exclude the presence of bugs.
Any bug can be reported via the GitHub issue tracker.
- •
- Homepage: <https://www.pgbarman.org/>
- •
- Documentation: <https://docs.pgbarman.org/>
- •
- Professional support:
<https://www.enterprisedb.com/>
Barman is the property of EnterpriseDB UK Limited and its code is distributed
under GNU General Public License v3.
© Copyright EnterpriseDB UK Limited 2011-2023
EnterpriseDB <
https://www.enterprisedb.com>.