Provided by: barman-cli-cloud_3.4.0-1_all bug

NAME

       barman-cloud-backup - Backup a PostgreSQL instance and stores it in the Cloud

SYNOPSIS

       barman-cloud-backup [OPTIONS] DESTINATION_URL SERVER_NAME

DESCRIPTION

       This  script  can  be used to perform a backup of a local PostgreSQL instance and ship the
       resulting tarball(s) to the Cloud.  Currently AWS S3, Azure Blob Storage and Google  Cloud
       Storage are supported.

       It requires read access to PGDATA and tablespaces (normally run as postgres user).  It can
       also be used as a hook script on a barman server, in which case it requires read access to
       the directory where barman backups are stored.

       If  the  arguments prefixed with --snapshot- are used, and snapshots are supported for the
       selected cloud provider, then the backup will be performed using snapshots  of  the  disks
       specified  using  --snapshot-disk arguments.  The backup label and backup metadata will be
       uploaded to the cloud object store.

       This script and Barman are  administration  tools  for  disaster  recovery  of  PostgreSQL
       servers written in Python and maintained by EnterpriseDB.

       IMPORTANT:  the  Cloud  upload  process  may fail if any file with a size greater than the
       configured  --max-archive-size  is  present  either  in  the  data  directory  or  in  any
       tablespaces.   However, PostgreSQL creates files with a maximum size of 1GB, and that size
       is always allowed, regardless of the max-archive-size parameter.

Usage

              usage: barman-cloud-backup [-V] [--help] [-v | -q] [-t]
                                         [--cloud-provider {aws-s3,azure-blob-storage,google-cloud-storage}]
                                         [--endpoint-url ENDPOINT_URL] [-P PROFILE]
                                         [--read-timeout READ_TIMEOUT]
                                         [--credential {azure-cli,managed-identity}]
                                         [-z | -j | --snappy] [-h HOST] [-p PORT] [-U USER]
                                         [--immediate-checkpoint] [-J JOBS]
                                         [-S MAX_ARCHIVE_SIZE] [-d DBNAME] [-n BACKUP_NAME]
                                         [--snapshot-instance SNAPSHOT_INSTANCE]
                                         [--snapshot-disk NAME]
                                         [--snapshot-zone SNAPSHOT_ZONE]
                                         [--snapshot-gcp-project SNAPSHOT_GCP_PROJECT]
                                         [--tags [TAGS [TAGS ...]]] [-e {AES256,aws:kms}]
                                         [--encryption-scope ENCRYPTION_SCOPE]
                                         destination_url server_name

              This script can be used to perform a backup of a local PostgreSQL instance and
              ship the resulting tarball(s) to the Cloud. Currently AWS S3, Azure Blob
              Storage and Google Cloud Storage are supported.

              positional arguments:
                destination_url       URL of the cloud destination, such as a bucket in AWS
                                      S3. For example: `s3://bucket/path/to/folder`.
                server_name           the name of the server as configured in Barman.

              optional arguments:
                -V, --version         show program's version number and exit
                --help                show this help message and exit
                -v, --verbose         increase output verbosity (e.g., -vv is more than -v)
                -q, --quiet           decrease output verbosity (e.g., -qq is less than -q)
                -t, --test            Test cloud connectivity and exit
                --cloud-provider {aws-s3,azure-blob-storage,google-cloud-storage}
                                      The cloud provider to use as a storage backend
                -z, --gzip            gzip-compress the WAL while uploading to the cloud
                -j, --bzip2           bzip2-compress the WAL while uploading to the cloud
                --snappy              snappy-compress the WAL while uploading to the cloud
                -h HOST, --host HOST  host or Unix socket for PostgreSQL connection
                                      (default: libpq settings)
                -p PORT, --port PORT  port for PostgreSQL connection (default: libpq
                                      settings)
                -U USER, --user USER  user name for PostgreSQL connection (default: libpq
                                      settings)
                --immediate-checkpoint
                                      forces the initial checkpoint to be done as quickly as
                                      possible
                -J JOBS, --jobs JOBS  number of subprocesses to upload data to cloud storage
                                      (default: 2)
                -S MAX_ARCHIVE_SIZE, --max-archive-size MAX_ARCHIVE_SIZE
                                      maximum size of an archive when uploading to cloud
                                      storage (default: 100GB)
                -d DBNAME, --dbname DBNAME
                                      Database name or conninfo string for Postgres
                                      connection (default: postgres)
                -n BACKUP_NAME, --name BACKUP_NAME
                                      a name which can be used to reference this backup in
                                      commands such as barman-cloud-restore and barman-
                                      cloud-backup-delete
                --snapshot-instance SNAPSHOT_INSTANCE
                                      Instance where the disks to be backed up as snapshots
                                      are attached
                --snapshot-disk NAME  Name of a disk from which snapshots should be taken
                --snapshot-zone SNAPSHOT_ZONE
                                      Zone of the disks from which snapshots should be taken
                --tags [TAGS [TAGS ...]]
                                      Tags to be added to all uploaded files in cloud
                                      storage

              Extra options for the aws-s3 cloud provider:
                --endpoint-url ENDPOINT_URL
                                      Override default S3 endpoint URL with the given one
                -P PROFILE, --profile PROFILE
                                      profile name (e.g. INI section in AWS credentials
                                      file)
                --read-timeout READ_TIMEOUT
                                      the time in seconds until a timeout is raised when
                                      waiting to read from a connection (defaults to 60
                                      seconds)
                -e {AES256,aws:kms}, --encryption {AES256,aws:kms}
                                      The encryption algorithm used when storing the
                                      uploaded data in S3. Allowed values:
                                      'AES256'|'aws:kms'.

              Extra options for the azure-blob-storage cloud provider:
                --credential {azure-cli,managed-identity}
                                      Optionally specify the type of credential to use when
                                      authenticating with Azure Blob Storage. If omitted
                                      then the credential will be obtained from the
                                      environment. If no credentials can be found in the
                                      environment then the default Azure authentication flow
                                      will be used
                --encryption-scope ENCRYPTION_SCOPE
                                      The name of an encryption scope defined in the Azure
                                      Blob Storage service which is to be used to encrypt
                                      the data in Azure

              Extra options for google-cloud-storage cloud provider:
                --snapshot-gcp-project SNAPSHOT_GCP_PROJECT
                                      GCP project under which disk snapshots should be stored

REFERENCES

       For Boto:

       • https://boto3.amazonaws.com/v1/documentation/api/latest/guide/configuration.html

       For AWS:

       • https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-getting-set-up.html

       • https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-getting-started.html.

       For Azure Blob Storage:

       • https://docs.microsoft.com/en-us/azure/storage/blobs/authorize-data-operations-cli#set-environment-variables-for-authorization-parameters

       • https://docs.microsoft.com/en-us/python/api/azure-storage-blob/?view=azure-python

       For libpq settings information:

       • https://www.postgresql.org/docs/current/libpq-envars.html

       For           Google           Cloud           Storage:           *           Credentials:
       https://cloud.google.com/docs/authentication/getting-started#setting_the_environment_variable

       Only authentication with GOOGLE_APPLICATION_CREDENTIALS env is supported at the moment.

DEPENDENCIES

       If using --cloud-provider=aws-s3:

       • boto3

       If using --cloud-provider=azure-blob-storage:

       • azure-storage-blob

       • azure-identity (optional, if you wish to use DefaultAzureCredential)

       If using --cloud-provider=google-cloud-storage * google-cloud-storage

       If using --cloud-provider=google-cloud-storage with snapshot backups

       • grpcio

       • google-cloud-compute

EXIT STATUS

       0      Success

       1      The backup was not successful

       2      The connection to the cloud provider failed

       3      There was an error in the command input

       Other non-zero codes
              Failure

SEE ALSO

       This script can be used in conjunction with post_backup_script or post_backup_retry_script
       to relay barman backups to cloud storage as follows:

              post_backup_retry_script = 'barman-cloud-backup [*OPTIONS*] *DESTINATION_URL* ${BARMAN_SERVER}'

       When running as a hook script, barman-cloud-backup will read the location  of  the  backup
       directory  and  the  backup  ID from BACKUP_DIR and BACKUP_ID environment variables set by
       barman.

BUGS

       Barman has been extensively tested, and is currently  being  used  in  several  production
       environments.  However, we cannot exclude the presence of bugs.

       Any bug can be reported via the GitHub issue tracker.

RESOURCES

       • Homepage: <https://www.pgbarman.org/>

       • Documentation: <https://docs.pgbarman.org/>

       • Professional support: <https://www.enterprisedb.com/>

COPYING

       Barman  is  the  property of EnterpriseDB UK Limited and its code is distributed under GNU
       General Public License v3.

       © Copyright EnterpriseDB UK Limited 2011-2023

AUTHORS

       EnterpriseDB <https://www.enterprisedb.com>.