Open mzwettler2 opened 4 months ago
Hi @mzwettler2 , if you want to restore your data back to a specific time (e.g. to three hours prior, specifically due to an issue with your data), this sounds like a Disaster Recovery (DR) scenario. More specifically, it sounds like you want a point-in-time restore (PITR), as discussed in the following docs:
I therefore recommend this as the best/safest way to meet your needs/use-case.
Hi @andrewlecuyer, thanks for your answer.
we have to be back online very quickly in the event of an error. the database is very large. a PITR would take too long. that's why we want to work with a lagging standby database, which is a very common use case. we also do this in classic database operation. unfortunately, the crunchy implementation does not currently enable it on kubernetes.
We have a simple configuration with 2 replicas (1 primary + 1 standby).
We configured the standby to run 3 hours behind the primary:
In case there is a logical problem on the primary (wrong data processing, misleaded application upgrade,…) we want to perform a manual failover to the standby, which still contains the old, correct data status. That means the standby should not apply any more WALs within the 3 hour residue in the event of a manual failover.
I could not find any working solution. Whatever I tried the standby first applied all outstanding WALs (which I don't want) and only then promoted the standby. Any idea how to get this working?