-
Notifications
You must be signed in to change notification settings - Fork 252
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Volume snapshots WAL file Issues when spinning up replicas #4488
Comments
If I use barman for backups, the replica spins up using pg_basebackup. and has no issues. |
Remember that this is a standby. So the same limitations of PostgreSQL apply here (see https://cloudnative-pg.io/documentation/current/backup/#backup-from-a-standby - the blue box). Is this server under workload? Have you tried running the backup a few minutes after the server is up (my fear is that you don't have yet the WAL file containing the checkpoint coming from the primary when you run that backup and when you create the replica). Can you please share the backup and the volume snapshot resources too? Thanks. |
Is there an existing issue already for this bug?
I have read the troubleshooting guide
I am running a supported version of CloudNativePG
Contact Details
No response
Version
1.23.0
What version of Kubernetes are you using?
1.27
What is your Kubernetes environment?
Cloud: Azure AKS
How did you install the operator?
YAML manifest
What happened?
Using (1.23.1 cnpg, postgres image 13.14-18). When creating a relica, increasing the instance count by 1, I first create an online volume snapshot. Taking an online volume snapshot appears to create an issue with the WAL files. The replica spins up using the volume snapshot, but then gets stuck processing the WAL file created after the backup finished.
Cluster resource
Relevant log output
Code of Conduct
The text was updated successfully, but these errors were encountered: