Real-world Use Case: Avoid Data Corruption by Switching from Passive Replication to Automated Failover and Recovery

Recently, a customer performing data validation identified some missing DR files and contacted Gravic for assistance.  Upon review, the customer identified a misconfigured application as the issue.

Application Failover Did Not Include All of the Customer’s Applications

The customer replicates data in a uni-directional Active / Passive Post-Disaster Recovery configuration using two HPE NonStop systems.

Figure 1 – Uni-directional Active / Passive Post-Disaster Recovery
Figure 1 – Uni-directional Active / Passive Post-Disaster Recovery

Every few months, the customer does failover testing. This architecture requires the customer to manually failover all of the required applications and switch replication in the opposite direction.

Figure 2 – Failover Testing
Figure 2 – Failover Testing

For some reason, one of the customer’s applications was not properly configured to failover and process transactions in the reverse direction.

Fixing Data Corruption

Upon further investigation and troubleshooting, when the application was running on the failover node, the customer noticed records changing on the old Production database. The data applied to the old Production database needed to be reapplied to the new Production database (on the failover node). In other words, the two nodes had switched roles, and one of the applications was pointing to the wrong database!

Figure 3 – Improper Failover Leads to Failover Faults
Figure 3 – Improper Failover Leads to Failover Faults

What Could Possibly Go Wrong?

The customer noticed and remediated this issue before serious data corruption occurred. They were lucky. Misconfigurations like this can happen in Active / Passive architectures and occur far more often than you may think.

Still, this use case highlights several important concepts:

  1. The importance of Data Validation
    — With frequent Shadowbase Compare runs, data discrepancies are flagged in a report, regardless of the source and target database polarities
  2. The dangerous nature of Failover Faults
    — Had the customer needed to failover to its backup, a portion of its data would be missing and have “disappeared” along with a “hung” application still making changes to the incorrect (former Production) database
  3. The importance of Automated Failover and Recovery Detection
    1. The customer utilized an Active / Passive replication architecture which requires a manual application switchover (reconnecting the app from \PROD -> \DR)
    2. They did not have warnings enabled to flag when data was being changed “on the wrong database” (called Reverse Replication Cut-off in HPE Shadowbase)
    3. The application switchover process was error-prone – the technologist performing the switchover tried their best but failed to include some of their applications in this process
    4. This approach resulted in most of their apps successfully switching over, but there were still some apps that were incorrectly accessing the old Production system
    5. If the customer had chosen a Shadowbase Sizzling-Hot-Takeover (SZT) architecture instead, the failover fault could have been detected based on user-defined conditions set in HPE Shadowbase
    6. In addition, in an SZT environment, this failover process could have been tested and verified ahead of time to reduce and effectively eliminate the failover fault risks

Key Takeaways

Business Continuity Continuum

This use case highlights the importance of the Business Continuity Continuum (see Figure 4). It may seem difficult to implement Continuous Availability, but sometimes it is easier than you think. Other times, there are work-a-rounds and customizations that can be implemented for speed and efficiency. With a little help and a different perspective, what may seem like “a mountain to climb” may actually only be a hill.

Regardless, it is never a bad time to reassess your BC strategy, and we are happy to assist in any way possible.  For reference, here is a throwback to a classic Gravic white paper: Choosing a Business Continuity Solution to Match Your Business Availability Requirements

Topics include:

    1. Redundancy
    2. The Components of Availability
    3. Application-Services Availability and the Recovery Time Objective (RTO)
    4. Data Availability and the Recovery Point Objective (RPO)
    5. Specifying Availability with RTO and RPO
    6. An Introduction to the Business Continuity Continuum
    7. Data Replication – The Fundamental Force Behind High and Continuous Availability
    8. An Overview of Data Replication Techniques
    9. Hardware versus Software Replication
    10. Asynchronous versus Synchronous Replication
    11. Uni-directional versus Bi-directional Replication
    12. Data Replication Provides Both High and Continuous Availability
    13. Eliminating Planned Downtime
    14. Choosing a Replication Method
    15. The HPE Shadowbase Data Replication Product Suite
    16. HPE Shadowbase Asynchronous Data Replication Engine

Thanks for reading. We hope you had a wonderful Holiday season, and we wish you luck in the New Year!

Author

  • Paden R. Holenstein

    Paden Holenstein works in marketing communications, business development, and branding for the HPE Shadowbase Products Group. He is a “NonStop Under 40” member, who attends and presents at regional TUG/RUG meetings and partner technical conferences. He writes extensively for the Shadowbase Products Group and is also responsible for the Shadowbase website. He graduated from West Chester University with a B.S. in Marketing and a minor in Information Technology. For more information, please see HPE Shadowbase Zero Downtime Migration (ZDM) or read the white paper, Using HPE Shadowbase Software to Eliminate Planned Downtime via Zero Downtime Migration.

Be the first to comment

Leave a Reply

Your email address will not be published.


*