I need a solution
Hi
I have a server cluster environment with VCS 6.0.2, 6 servers constitute the cluster, I/O fencing is configured with 3 coordinator disks. If I cold boot one of server, I found CFS access on other running server is blocking in a period.
I found CFS access starts being blocked when below logs appear in /var/log/messages
LLT INFO V-14-1-10205 link 0 (eth6.109) node 0 in trouble LLT INFO V-14-1-10205 link 1 (eth7.110) node 0 in trouble
And access allowing when below logs appear in /var/log/messages
vxfs: msgcnt 8 Phase 2 - /dev/vx/dsk/filedg/filevol - Buffer reads allowed. vxfs: msgcnt 9 Phase 9 - /dev/vx/dsk/filedg/filevol - Set Primary nodeid to 2 vxglm INFO V-42-106 GLM recovery complete, gen f59d30, mbr 2c/0/0/0 vxglm INFO V-42-107 times: skew 2673 ms, remaster 78 ms, completion 40 ms
I think the CFS access blocking is for data protection, but as my observation, CFS access blocking may continue 10+ seconds on running servers, so my questions are:
1. Is this the correct behaviour for VCS to block CFS access 10+ seconds?
2. Why not start CFS access blocking after heartbeat link being expired and before racing coordinator disks.
Thanks in advance!