Hello All,
I hope I am entering this in properly.
I have two 2-node clusters both under the same VC organization (Hope thats a correct term), they both have their own datastore (NETAPP) and everything works fine. I want to combine the two into one 4-node cluster. I have given all 4 nodes the same access to both datastores and have mounted them with no problem. Both Clusters do have DRS and HA enabled and it is working fine in each. The machines are identical with the exception of one node, it has 24Gig of RAM as opposed to the others having 32Gb. (8 more Gigs are on the way)
Leaving HA and DRS enabled in the destination cluster, I moved one of the nodes form the source cluster to the destination cluster with no problem or errors....however...almost immediately the cluster wanted to balance the load and vMotion a VM to the new node...thats where the trouble began. vMotion ran up to about 97% then failed and it brought the VM down with it, I could not remove it from inventory and re-add or anything, I bascially had to delete it and restore from a NETApp snapshot. I did try removing HA nad DRS from the cluster and re-adding but the result was the same. I did validate that VMKPing was giving good responses and the all ESX hosts had full access to the datastores.
I have since moved the ESX host back to the original cluster where everything works fine.
Am I doing something wrong, is there a "Best Practice" on moving ESX hosts from one cluster to another?...seems like it should be pretty straight forward.
Any ideas?
Thanks in advance.