Cloud Sql Backup Save Abandoned

A script for backing up GCP Cloud SQL instances

Project README

cloud_sql_backup

CircleCI

This is a bash script that can be used for backing up GCP Cloud SQL instances. It works by restoring an existing GCP managed backup to a new ephemeral db instance, and exporting an SQL dump from there into a GCS bucket.

Install

The cloud_sql_backup.sh script can be used on its own, or in a Docker image we've prepared.

Pre-requisites

  1. An automated or on-demand GCP managed backup. Go here for help enabling.

  2. The host running the script must have these tools installed:

    • cut
    • date
    • gcloud
    • head
    • jq
    • sed
    • tr

Configuring

The script requires some environment variables in order to function correctly:

Name Description Example
DB_NAME (colon separated) The database name(s) that'll be exported to GCS "my-db" or "my-db-1:my-db-2"
INSTANCE_CPU vCPUs of the ephemeral instance "4"
INSTANCE_ENV Name of environment the backup process is running in. It's used in the ephemeral instance name "nonprod"
INSTANCE_MEM Memory of instance (multiple of 256 MiB, min 3840MiB) "7680MiB"
INSTANCE_NAME_PREFIX Prefix to add to the start of instance name "my-backup"
INSTANCE_REGION Instance region "europe-west1"
INSTANCE_STORAGE_TYPE SSD (default) or HDD "SSD"
PROJECT The GCP project "my-gcp-project"
SA_KEY_FILEPATH The path to the GCP service account's .json key "/secrets/gcp/backup-key.json"
SOURCE_BACKUP_INSTANCE Name of instance that you want backing up "uat-db"
TARGET_BACKUP_BUCKET URI of GCS bucket that the backup file will get written to "gs://my-gcs-bucket"

Optionally, you can set some env vars for metrics:

Name Description Example
DATADOG_API_KEY The Datadog API key "57dsfjho34kj3h4oknk4jj55"
TEAM Team name, used for tagging metrics "my-team"

The following metrics should appear in Datadog (depending on the success of subsequent backup attempts):

  • cloud.sql.backup.started.count
  • cloud.sql.backup.success.count
  • cloud.sql.backup.failure.count
  • cloud.sql.backup.cleanup.count

All metrics are of type: count, and a value of 1 is posted when each phase of the backup process has been reached.

Why is this required?

GCP provide their own automated daily backup process, which is very easy to enable, so why is this required?

According to GCP, "All data on an instance, including backups, is permanently lost when that instance is deleted". They recommend exporting your data to cloud storage for longer term persistence. That export process, however, can have a serious impact on the performance of your database while it's taking place.

This script solves these problems by:

  • Creating a database instance from the latest automated GCP-managed backup
  • Giving the database instance the required permissions to write to a GCS bucket
  • Exporting to an SQL dump file in a GCS bucket

The ephemeral database instance and elevated permissions are then deleted/removed.

What permissions are required?

When this script runs using a GCP service account, it'll need a specific set of permissions in order to elevate the database instance service account's permissions, and check that the SQL dump file has appeared in the GCS bucket:

  permissions = [
    "cloudsql.backupRuns.get",
    "cloudsql.backupRuns.list",
    "cloudsql.instances.create",
    "cloudsql.instances.delete",
    "cloudsql.instances.export",
    "cloudsql.instances.get",
    "cloudsql.instances.restoreBackup",
    "storage.buckets.get",
    "storage.buckets.getIamPolicy",
    "storage.buckets.setIamPolicy",
    "storage.buckets.update",
    "storage.objects.get"
  ]

The service account must have OWNER-role access to the TARGET_BACKUP_BUCKET. GCP creates a temporary service account during the export process, which needs to be able to write to your GCS bucket. The service account you provide needs enough permission to elevate the temporary service account's permissions on the bucket (they get revoked again after being used).

Notes

Permissive Role

The least privileged role for the script to succeed, as detailed here, is permissive. The "cloudsql.instances.delete" permission alone will allow anyone with the key, for the GCP service account you use, to delete any Cloud SQL instance in its GCP project.

Instance Deletion

The final step in cloud_sql_backup.sh is to delete the ephemeral db instance that's been used to create the SQL dump. There's a hard-coded check at this point in the script to only perform the instance deletion when the instance name contains the string "backup".

Naming Format

To prevent a race condition, the script creates a name, to be used for both the ephemeral instance and the S3 SQL dump object, that's suffixed by a 5 character random string:

name = <instance_name_prefix>-<instance_env>-<timestamp>-<gcp_managed_backup_id>-<random_string>
  • timestamp is obtained by executing date +%Y%m%d%H%M%S
  • instance_name_prefix and instance_env are obtained from env vars
  • gcp_managed_backup_id is the ID of the latest GCP managed backup
  • random_string is the value of LC_ALL=C tr -dc 'a-z0-9' </dev/urandom | head -c 5

The .gz file that gets written to GCS will also have the db name as the last element before the file extension.

Backup candidate

The ID of the latest successful GCP managed backup is obtained using:

BACKUP_DATA=$(gcloud sql backups list \
  --instance "$SOURCE_BACKUP_INSTANCE" \
  --filter STATUS=SUCCESSFUL \
  --limit 1 | sed 1,1d | tr -s ' ')
BACKUP_ID=$(echo "$BACKUP_DATA" | cut -d ' ' -f 1)

Whilst it's recommended to monitor for failed/successful cloud_sql_backup.sh script executions, this can't be relied upon to report freshness of data. For example, your GCP managed backups may start failing, but the cloud_sql_backup.sh script will keep succeeding, but with out-of-date data (based on the last successful GCP managed backup).

Completion Check

The penultimate task of the cloud_sql_backup.sh script is to poll GCS using gsutil to verify the object (SQL dump) has arrived in GCS as expected. This has to be performed out-of-band of the SQL dump process, as the dump is an operation that's triggered on the ephemeral db instance (using gcloud sql export sql).

Contributions

Contributions are very welcome. Please branch or fork, and submit a PR.

PRs from branches will result in an e2e test being run in CircleCI, which can sometimes take around 15mins. Commits from forked PRs should be made onto a branch in this repo, and another PR opened so the e2e test can run.

Open Source Agenda is not affiliated with "Cloud Sql Backup" Project. README Source: ovotech/cloud_sql_backup
Stars
30
Open Issues
4
Last Commit
1 year ago
License

Open Source Agenda Badge

Open Source Agenda Rating