This page describes how to troubleshoot and fix replication lag for Cloud SQL read replicas.
Overview
Cloud SQL read replicas use PostgreSQL streaming replication. Changes are written to Write-Ahead Log (WAL) in the primary instance. The WAL sender sends the WAL to the WAL receiver in the replica, where they are applied.Replication lag can happen in a few scenarios, such as:
- The primary instance can't send the changes fast enough to the replica.
- The replica can't receive the changes quickly enough.
- The replica can't apply the changes quickly enough.
network_lag metric.
The third is observed via the replica_lag metric. High replica_lag means that
the replica can't apply replication changes fast enough. The total lag can be
observed via replica_byte_lag metric, which has labels to indicate further
details. These metrics are described in the Monitor replication lag
section below.
Ensure replica is adequately provisioned
A replica instance that is smaller than the primary instance (for example, with less vCPUs and memory) can experience replication lag. A smaller replica might also have different default configuration flags compared to a larger primary instance. We recommend that the replica instance is at least as large as the primary instance to have enough resources to handle the replication load.
High CPU utilization on the replica can also cause replication lag. If the replica's CPU utilization is high (for example, over 90%), consider increasing the replica's CPU capacity.
You can use theSHOW ALL command to see replica and primary instance
configuration and compare them for differences.
Optimize queries and schema
This section suggests some common query and schema optimizations you can make to improve replication performance.
Long-running queries in the read replica
Long-running queries in the replica might block replication for Cloud SQL.
This can happen when replication is trying to apply changes (like from a
VACUUM operation) to rows that are being read by a query on the replica.
You might want to have separate replicas for online transaction processing (OLTP) and online analytical processing (OLAP) purposes and only send long-running queries to the OLAP replica.
To help address replication delays or blockages caused by long-running transactions, we recommend the following:
-
Adjust standby delay flags. The
max_standby_archive_delayandmax_standby_streaming_delayflags control how long a replica will wait before canceling standby queries that conflict with replication. Reasonable values are often around 30 to 60 seconds. You can check thepg_stat_database_conflictsview for insights about query conflicts. -
Enable the
hot_standby_feedbackflag. Setting thehot_standby_feedbackflag toonin the replica can help by delaying vacuum operations on the primary. However, this can cause table bloat on the primary, so it's a trade-off.
Review PostgreSQL documentation for more information.
High network lag
High network lag indicates that WAL records are not being sent by the primary or received by the replica fast enough. This can be caused by the following:
- Cross-region replication. Replicating between different regions can introduce higher network latency.
- High primary CPU utilization. If the primary's CPU is over 90%, the WAL sender process might not get enough CPU time. Consider reducing load on the primary or increasing its CPU.
- High replica CPU utilization. If the replica's CPU is over 90%, the WAL receiver process might not get enough CPU time. Consider reducing load on the replica or increasing its CPU.
- Network bandwidth issues or disk I/O bottlenecks. A closer region or
a higher throughput disk configuration might help. Consider modifying the
wal_compressionflag value in the primary instance to help reduce cross-region traffic.
cloudsql.googleapis.com/database/replication/network_lag metric.
This metric has a maximum limit of 25 seconds, even if the actual lag is higher.
This network_lag metric is similar to the cloudsql.googleapis.com/database/postgresql/replication/replica_byte_lag
metric which measures the sent_location lag in terms of bytes indicated by its replica_lag_type label.
Exclusive locks due to DDL
Data definition language (DDL) commands, such as ALTER TABLE and
CREATE INDEX, can cause replication lag in the replica due to
exclusive locks. To avoid lock contention, consider scheduling DDL execution
during times when the query load is lower on the replicas.
Overloaded replica
If a read replica is receiving too many queries, replication could be blocked. Consider splitting the reads among multiple replicas to reduce the load on each one.
To avoid query spikes, consider throttling replica read queries in your application logic or in a proxy layer if you use one.
If there are spikes of activity on the primary instance, consider spreading out updates.
Monolithic primary database
Consider sharding the primary database vertically (or horizontally) to prevent one or more lagging tables from holding back all the other tables.
Monitor replication lag
You can use the replica_lag and network_lag metrics to monitor replication
lag and identify whether the cause of the lag is in the primary database,
the network, or the replica.
| Metric | Description |
|---|---|
| Replication lag ( cloudsql.googleapis.com) |
The number of seconds that the replica's state is lagging behind the state of the primary instance. This is the difference between the current time and the original timestamp at which the primary database committed the transaction that is currently being applied on the replica. In particular, writes might be counted as lagging even if they have been received by the replica, if the replica hasn't yet applied the write to the database. This metric is calculated using |
| Lag bytes ( cloudsql.googleapis.com) |
The amount of bytes by which the replica's state is lagging behind the
state of the primary database.
|
| Network lag ( cloudsql.googleapis.com) |
The amount of time in, seconds that it takes from commit in the primary database to reach the WAL receiver in the replica. If the |
Verify replication
To verify that replication is working, run the following statement against the replica: select status, last_msg_receipt_time from pg_stat_wal_receiver;
If replication is happening, you see the status streaming and a recent
last_msg_receipt_time:
postgres=> select status, last_msg_receipt_time from pg_stat_wal_receiver;
status | last_msg_receipt_time
-----------+-------------------------------
streaming | 2020-01-21 20:19:51.461535+00
(1 row)
If replication is not happening, an empty result is returned:
postgres=> select status, last_msg_receipt_time from pg_stat_wal_receiver;
status | last_msg_receipt_time
--------+-----------------------
(0 rows)