I know there are already plenty of topics (I am still in the process of reading them all) but I would like to get some more technical experience with hardware clustering and webMethods 6.1.
For now, we have available:
- 1 to many IS 6.1 SP2 (proxy) as frontend, and 1 to many IS 6.1 SP2 (executing) as backend. Backend is using Scheduler, Monitor, Broker and PRT.
- 1 Broker 6.1 SP3
- 1 LTM1500 (F5, formerly known as BigIP) hardware load balancer
All this running on a Linux Red Hat 3 box.
We want to implement high availability that would prevent single point of failure except for:
- database (considerered HA)
- NFS file system (considered HA as well)
How should we implement the HA option, given these tools available to support :
- LB of frontend (optional but recommended)
- LB of backend (required)
- LB of broker (I know only OS clustering is supported: but how to implement it?)
Questions about this:
- what pros/cons about using NFS shares to share data between clustered IS?
- how to keep scheduled services synchronized between the IS in the cluster?
- does hardware clustering impacts functional behavior of PRT/BPMs in Monitor? More generally Broker impacts associated to clustering?
- should we share DB or file repo between load balanced IS? (in the backend)
- any known impacts you ran into, that we could learn from before experiencing the same issues?
If you have some evolved scripts to configure the BigIP LLB, let share it to us as well. For instance, how do you make sure the IS is “healthy” (at a higher iso layer than 3)? Is it thinkable to implement a routine check that would make sure IS is up, Broker is up, File system is available, etc, to identify the readiest IS available to take incoming requests.
Thanks for sharing your experiences!