All, the problem we were having was directly related to IBM equipment running Microsoft Cluster. Our other clusters were failing over fine. This is IBM specific and the firmware upgrade to 8.x fixes the problem. They have noted this in their lab in raleigh.
The nodes can ping each other on the public and private networks. The nodes on the private are setup as 10.1.1.1 and .2 with a subnet of 255.0.0.0.
The subnet on the public is 255.255.255.0
We finally figured out what was going on with our setup. We had the 300Gs cabled to a FastT 500 storage array from IBM. After 2 weeks of pain and suffering, we finally got them to say yes to a firmware upgrade on the storage Fastt 500.
Upgraded the firmware and the cluster was functioning and...
This sounds exactly like the problem we are having. Running cluster on a 300G NAS device with dual channel qlogic fibres. When we pull both fibre connections, cluster dies.
Running MCSC on an IBM NAS 300G boxes with RDAC installed. We pull the primary fibre and RDAC fails over to the secondary path. When the 2nd path is pulled, the cluster fails.
Same scenario if we do a hard power down on the box.
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.