a9s-pg Major Upgrade
Up until anynines-deployment version v51.0.0
a9s-pg uses the a9s PostgreSQL 11 BOSH release which requires
the Bionic stemcell.
This document describes the steps necessary to safely upgrade a9s-pg to a9s PostgreSQL 15 using the
Jammy stemcell.
We recommend executing a test run in a development environment, with a clone of the production environment, to better prepare for the upgrade since there is a downtime inherent to it and the data stored in the cluster, which varies from case to case.
The upgrade to a9s PostgreSQL 15 is only supported when upgrading from a9s PostgreSQL 11 version 3.0.0
or later.
Make sure the cluster is healthy before executing the upgrade.
Description
The upgrade procedure takes place in the pre-start step of the bootstrap node update with the following steps:
- Initialize the new data directory (
/var/vcap/store/postgresql15
) withinitdb
. - Ensure there is a valid primary.
- Discard the current data in the old data dir (
/var/vcap/store/postgresql11
). - Clone the data from the current primary using pg_basebackup to ensure the node is up-to-date.
- Block all nodes in the cluster. See Downtime for more information.
- Prepare the data directory for upgrade.
- Validate and upgrade the data using pg_upgrade with
--link
. - Reindex all databases. See Reindex All Databases for more information.
Then, normal initialization of the node as primary proceeds, as the node has just been bootstrapped, initializing a new cluster in the new version containing all the data. Subsequent nodes are upgraded, discarding the current data and initializing the node following the new primary in the bootstrap node.
The steps described above require both PostgreSQL binary versions, so the new a9s PostgreSQL BOSH release contains the PostgreSQL binaries from the old version.
When BOSH is updating the bootstrap node of the a9s-pg deployment with the new a9s PostgreSQL 15, the pre-start
script
detects that a major upgrade is happening based on the PG_VERSION
file of the old data directory
(/var/vcap/store/postgresql11
), the existence of this file is necessary for the success of the upgrade procedure.
Any attempt to upgrade the cluster with an empty bootstrap node will cause data loss, as the upgrade procedure will not be executed.
Downtime
The in-place major version upgrade procedure has an inherent downtime due to:
- Bootstraping a new cluster: after cloning finishes on the bootstrap node, any data written to the primary of the cluster is not replicated to the bootstrap node and is discarded during the upgrade. The part of the cluster in the old version is blocked after the bootstrap node finishes cloning to avoid losing data, and connections are only accepted when the bootstrap node is ready to accept connections again.
- The way
pg_upgrade
manipulates the local data to provide compatibility with the new version: this tool requires the PostgreSQL process to be stopped. The number of databases, amount of data, and disk throughput affect how long this operation takes. - Stemcell upgrade: the operating system update affects glibc and potentially corrupts the indexes leading to data corruption and data loss, therefore requiring a reindexing operation to be executed on all existing databases. The number of databases, indexes, rows in tables, and the resources available in the machine affect how long this operation takes. See Reindex all databases for more information.
- Not enough quorum when bootstrapping the cluster in the new version: a9s PostgreSQL comes with PgGuard, a component that watches the cluster and blocks unhealthy nodes to prevent data loss. When the bootstrap node finishes upgrading, it has no replica, and to avoid exposing new changes to a cluster where new data is not redundant, PgGuard blocks the bootstrap node until the second node starts following the new primary. This behavior can be configured. See Unblock Bootstrap Node on Major Upgrade for more information.
We recommend measuring the downtime while staging the upgrade so you can prepare a more accurate maintenance window.
Preparing to Upgrade
Some steps are necessary to ensure a safe upgrade, and the possibility to recover in case of a disaster. This section describes some aspects to take into consideration.
Increase the Persistent Disk
During the execution of pg_upgrade
, data is copied from the old data directory (/var/vcap/store/postgresql11
) to the
new data directory (/var/vcap/store/postgresql15
) using hard links. Which should save a lot of space usage during the
upgrade. However, we recommend watching the storage usage when validating the upgrade in the staging environment paying
special attention to the a9s Parachute limit.
Create a Backup
Before going ahead with the upgrade, we recommend creating a backup in case disaster recovery is necessary.
Using the a9s Backup Framework
Please, ensure that your a9s-pg deployment contains a well configured a9s Backup Agent, and therefore does not work if
the ops/a9s-pg-without-backup.yml
ops file is applied.
Before triggering backup you need to check that it was configured correctly. If we look at the manifest of the a9s-pg,
we should see that we installed backup-agent
and postgresql-backup-endpoint
:
After checking that everything fine we can trigger back up. The instruction of how to do that you can find a9s-pg Backup.
Even though a backup of the a9s PostgreSQL 11 backup is not ensured to be compatible with the a9s PostgreSQL 15 version, it is possible to reset the cluster, roll back to the old version, and restore a backup. To know more about this, see the Disaster Recovery section below.
Using pg_dumpall
(Manual Backup)
Another option if to create a local dump file with the changes to be restored in case of a disaster with pg_dumpall by using the following command:
$ pg_dumpall --clean --oids --no-password --if-exists --quote-all-identifiers --host=<primary-addr> --port=5432 \
--username=admin_user -f a9s-pg.sql
The password for this user is on the CredHub variable a9s_pg_admin_password
. It should be under the path
/<bosh-directory-name>/a9s-pg/a9s_pg_admin_password
.
If the password does not seem to be working, try to set the PGPASSWORD
environment variable.
Even though a backup of the a9s PostgreSQL 11 backup is not compatible with the a9s PostgreSQL 15 version, it is possible to reset the cluster, roll back to the old version, and restore a backup. To know more about this, see the Disaster Recovery section below.
Customize The Upgrade
Unblock Bootstrap Node on Major Upgrade
In this section, we explain how to configure PgGuard to alter its behavior and unblock the bootstrap to start accepting connections as soon as the node starts without waiting for a quorum, therefore decreasing the downtime.
Enabling this feature, however, exposes any data written until the second node starts replicating from the bootstrap node to a vulnerable scenario as there is only one copy of the new changes, which presents a risk of data loss if data in the bootstrap node is lost due to any outage.
PgGuard can be configured to unblock the bootstrap node as soon as it starts with the following Ops File:
- type: replace
path: /instance_groups/name=pg/jobs/name=postgresql-ha/properties/postgresql-info-webservice/pg_guard?/unblock_bootstrap_major_upgrade?
value: true
The default value of the postgresql-info-webservice.pg_guard.unblock_bootstrap_major_upgrade
is false
, and node
are unblocked only when there is enough standby nodes following the primary, which is the behavior described in the
Description section.
Reindex All Databases
During the upgrade of a9s-pg to a9s PostgreSQL 15, the stemcell will be upgraded as well, and the cluster will start
using the Ubuntu Jammy Stemcell, with a newer version of the operating system. This version comes with glibc version
2.28
installed, which brings a collation change that can mess with the indexes of the databases created in the
previous version. See Locale data changes for more information.
The major upgrade procedure includes a step to automatically reindex all databases in a9s-pg to address this issue.
This step happens after the data is upgraded. The PostgreSQL process starts with a special configuration tailored for
the execution of the reindexdb command, listening only on
localhost
.
It is possible to skip the reindex step and decrease the downtime with the following Ops File:
- type: replace
path: /instance_groups/name=pg/jobs/name=postgresql-ha/properties/postgresql-info-webservice/major_upgrade?/reindex?/enabled?
value: false
We strongly adivice not to skip this step, as skipping it can lead to data corruption.
During this step, the reindex command is executed with the --jobs
specified as the same number of CPUs available in
the machine.
The PostgreSQL process is started with the following configuration by default:
max_connections
:(4 * number of CPUs)
. Default minimum value for this property is128
.max_worker_processes
: number of CPUs.max_parallel_workers
: number of CPUs.max_parallel_maintenance_workers
: number of CPUs.max_parallel_workers_per_gather
:(number of CPUs / 4)
. Default minumum value for this property is2
.maintenance_work_mem
:(50% of the total memory / (max_worker_processes * max_parallel_workers_per_gather))
. Default minimum value for this property is64MB
.
It is possible to override these values and provide any PostgreSQL configuration with the following Ops File:
- type: replace
path: /instance_groups/name=pg/jobs/name=postgresql-ha/properties/postgresql-info-webservice/major_upgrade?/reindex?/pg_ctl_start_opts?
value:
max_connections: 128
max_worker_processes: 64
max_parallel_workers: 64
max_parallel_maintenance_workers: 64
max_parallel_workers_per_gather: 4
maintenance_work_mem: 128MB
Any configuration supported by the --options
parameter of the pg_ctl
command under the -c
config can be
overwritten extending the Ops File above.
The optimal execution of the reindexdb command depends on the amount of data stored, so it makes sense to observe the resource usage and attempt to change these values to decrease the reindex operation time, and therefore decrease the downtime itself.
Run Upgrade
The upgrade takes place in main two steps when deploying the new version of a9s-pg BOSH deployment:
Update the a9s-pg deployment to most recent a9s PostgreSQL 11 version. This is currently the default version used in the
a9s-pg/a9s-pg.yml
manifest of the most recent anynines-deployment version. See Deploy a9s-pg for more information.Apply the a9s PostgreSQL 15 BOSH release to the a9s-pg deployment by deploying it again while applying the
a9s-pg/ops/upgrade-to-postgresql15.yml
Ops File. The major upgrade will happen during this step. See the command below for an example:
$ bosh deploy -d a9s-pg a9s-pg/a9s-pg.yml -l config/iaas-config.yml -o a9s-pg/ops/upgrade-to-postgresql15.yml
After applying the upgrade with success, the a9s-pg cluster will be ready for use. If it fails, please refer to the Disaster Recovery section below.
Disaster Recovery
In case the upgrade procedure fails, the cluster can be recovered by:
- Cleanup all nodes in the cluster:
$ monit stop all
$ rm -rf /var/vcap/store/postgresql15*
If the update fails while updating the bootstrap node, you can execute the cleanup in the bootstrap node only.
Deploy the previous version of your a9s-pg manifest, to roll back the version to 11.
If data is lost for any reason, restore the backup taken in the begin of the upgrade, and try to upgrade to the new version again as described on Run Upgrade section.
Restore a Backup
Using the a9s Backup Framework
See the Restore a Backup documentation for further information.
Restore a Manual Backup
Restore the local dump file created on the Using pg_dumpall
Manual Backup with the
command below:
$ psql --host=<primary-addr> --username=admin_user --port=5432 --dbname=postgres < a9s-pg.sql
It should be ok to see some errors related to the repmgr database when restoring the dump file since the repmgrd
is
using the database and already with the correct information. But in case of any issue, see the
a9s PostgreSQL Cluster Recovery
documentation.