Quantcast
Channel: Symantec Connect - Storage and Clustering - Discussions
Viewing all 543 articles
Browse latest View live

diskgroup resource failed to online in a VCS environment with "failed to import with groupreserve option as scsi3pr is disabled"

$
0
0
I need a solution

Hi,

I have a two node cluster. I am trying to create a simple service group with VxFS on VxVM. I have followed the steps outlined in http://www.symantec.com/business/support/index?page=content&id=TECH20539

When I try to bring the service group online, the diskgroup would fail to import with "groupreserve option as scsi3pr is disabled". I've checked the diskgroup for scsi3-pr compliance with the vxfentsthdw utility. It is passing as shown below.

 

Can you please tell me how can I enable scsi3pr on the diskgroup so it can online successfully.

 

Thanks,

David

 

 

 

################### Diskgroup  Service online Error ####################################

Jun 27 15:56:36 sca-m532-504-gd9 Had[3429]: [ID 702911 daemon.notice] VCS ERROR V-16-10001-1095 (sca-m532-504-gd10) DiskGroup:data_dg:online:Diskgroup [datadg] failed to import with groupreserve option as scsi3pr is disabled.
Jun 27 15:58:37 sca-m532-504-gd9 Had[3429]: [ID 702911 daemon.notice] VCS ERROR V-16-2-13066 (sca-m532-504-gd10) Agent is calling clean for resource(data_dg) because the resource is not up even after online completed.
Jun 27 15:58:39 sca-m532-504-gd9 Had[3429]: [ID 702911 daemon.notice] VCS ERROR V-16-10001-1095 (sca-m532-504-gd10) DiskGroup:data_dg:online:Diskgroup [datadg] failed to import with groupreserve option as scsi3pr is disabled.
Jun 27 16:00:40 sca-m532-504-gd9 Had[3429]: [ID 702911 daemon.notice] VCS ERROR V-16-2-13066 (sca-m532-504-gd10) Agent is calling clean for resource(data_dg) because the resource is not up even after online completed.
Jun 27 16:00:42 sca-m532-504-gd9 Had[3429]: [ID 702911 daemon.notice] VCS ERROR V-16-1-54031 Resource data_dg (Owner: Unspecified, Group: datagrp) is FAULTED on sys sca-m532-504-gd10
Jun 27 16:00:42 sca-m532-504-gd9 Had[3429]: [ID 702911 daemon.notice] VCS ERROR V-16-1-10205 Group datagrp is faulted on system sca-m532-504-gd10
 

################# SCSI3-PR Test of Diskgroup ######################################

sca-m532-504-gd9:/root
[rootD:]825> /opt/VRTS/bin/vxfentsthdw -n -c datadg          

Veritas vxfentsthdw version 6.0.3 Solaris

The utility vxfentsthdw works on the two nodes of the cluster.
The utility verifies that the shared storage one intends to use is
configured to support I/O fencing.  It issues a series of vxfenadm
commands to setup SCSI-3 registrations on the disk, verifies the
registrations on the disk, and removes the registrations from the disk.

The logfile generated for vxfentsthdw is /var/VRTSvcs/log/vxfen/vxfentsthdw.log.10882

Enter the first node of the cluster:
sca-m532-504-gd9
Enter the second node of the cluster:
sca-m532-504-gd10

********************************************

Testing sca-m532-504-gd9 /dev/vx/rdmp/zfs_71200_3 sca-m532-504-gd10 /dev/vx/rdmp/zfs_71200_3

Evaluate the disk before testing  ........................ No Pre-existing keys
RegisterIgnoreKeys on disk /dev/vx/rdmp/zfs_71200_3 from node sca-m532-504-gd9  Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_3 on node sca-m532-504-gd9  Passed
RegisterIgnoreKeys on disk /dev/vx/rdmp/zfs_71200_3 from node sca-m532-504-gd10  Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_3 on node sca-m532-504-gd9  Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_3 on node sca-m532-504-gd10  Passed
Preempt and abort key KeyA using key KeyB on node sca-m532-504-gd10 .... Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_3 on node sca-m532-504-gd9  Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_3 on node sca-m532-504-gd10  Passed
Remove key KeyB on node sca-m532-504-gd10 .............................. Passed
Check to verify there are no keys from node sca-m532-504-gd9 ........... Passed

ALL tests on the disk /dev/vx/rdmp/zfs_71200_3 have PASSED.
The disk is now ready to be configured for I/O Fencing on node sca-m532-504-gd9
as a COORDINATOR DISK.

ALL tests on the disk /dev/vx/rdmp/zfs_71200_3 have PASSED.
The disk is now ready to be configured for I/O Fencing on node sca-m532-504-gd10
as a COORDINATOR DISK.

********************************************

Testing sca-m532-504-gd9 /dev/vx/rdmp/zfs_71200_4 sca-m532-504-gd10 /dev/vx/rdmp/zfs_71200_4

Evaluate the disk before testing  ........................ No Pre-existing keys
RegisterIgnoreKeys on disk /dev/vx/rdmp/zfs_71200_4 from node sca-m532-504-gd9  Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_4 on node sca-m532-504-gd9  Passed
RegisterIgnoreKeys on disk /dev/vx/rdmp/zfs_71200_4 from node sca-m532-504-gd10  Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_4 on node sca-m532-504-gd9  Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_4 on node sca-m532-504-gd10  Passed
Preempt and abort key KeyA using key KeyB on node sca-m532-504-gd10 .... Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_4 on node sca-m532-504-gd9  Passed
Verify registrations for disk /dev/vx/rdmp/zfs_71200_4 on node sca-m532-504-gd10  Passed
Remove key KeyB on node sca-m532-504-gd10 .............................. Passed
Check to verify there are no keys from node sca-m532-504-gd9 ........... Passed

ALL tests on the disk /dev/vx/rdmp/zfs_71200_4 have PASSED.
The disk is now ready to be configured for I/O Fencing on node sca-m532-504-gd9
as a COORDINATOR DISK.

ALL tests on the disk /dev/vx/rdmp/zfs_71200_4 have PASSED.
The disk is now ready to be configured for I/O Fencing on node sca-m532-504-gd10
as a COORDINATOR DISK.

Removing test keys and temporary files, if any...

sca-m532-504-gd9:/root
[rootD:]826>

1372406385

hatype parameters

$
0
0
I need a solution

Where I can find guidelines of how HA type attributes are related to each other?

I am particulaly interesting in how can I setup timeouts in a right way. I would like to know if I change say, AgentStartTimeout, to 120 sec, which other attributes need to be changed. The problem I have is that JBoss agent I built times out the startup because it takes JBoss 1 min 30 sec to start.

Reards

Leonid

1372882603

VCS is trying to bring online SG on node B even though it is online on node A

$
0
0
I need a solution

I have two nodes VCS on solaris. Two SGs are configured as dependet:

nssitdb01-zone_sg is parralel, Autostart is on on both nodes. Failover service group dbhost-app_sg is dependet on nssitdb01-zone_sg, Autostart is on on both nodes.

After clean reboot (init 6) of node B (sirius) VCS started nssitdb01-zone_sg and then is trying to bring online dbhost-app_sg too, even though it is online on node A(arcturus)!

Hier is some info from engine_A.log:

2013/07/05 10:16:13 VCS NOTICE V-16-1-10438 Group nssitdb01-zone_sg has been probed on system sirius
2013/07/05 10:16:13 VCS NOTICE V-16-1-10442 Initiating auto-start online of group nssitdb01-zone_sg on system sirius
2013/07/05 10:16:33 VCS NOTICE V-16-1-10447 Group nssitdb01-zone_sg is online on system sirius
2013/07/05 10:16:33 VCS WARNING V-16-1-50045 Initiating online of parent group dbhost-app_sg, PM will select the best node

2013/07/05 10:16:33 VCS INFO V-16-1-10493 Evaluating sirius as potential target node for group dbhost-app_sg
2013/07/05 10:16:33 VCS INFO V-16-1-10162 Group dbhost-app_sg has not been fully probed on system sirius
2013/07/05 10:16:33 VCS INFO V-16-1-10493 Evaluating arcturus as potential target node for group dbhost-app_sg
2013/07/05 10:16:33 VCS INFO V-16-1-50010 Group dbhost-app_sg is online or faulted on system arcturus
2013/07/05 10:16:53 VCS NOTICE V-16-1-10438 Group dbhost-app_sg has been probed on system sirius
2013/07/05 10:16:53 VCS INFO V-16-1-50007 Initiating auto-start online of group dbhost-app_sg
2013/07/05 10:16:53 VCS INFO V-16-1-10493 Evaluating arcturus as potential target node for group dbhost-app_sg
2013/07/05 10:16:53 VCS INFO V-16-1-50010 Group dbhost-app_sg is online or faulted on system arcturus
2013/07/05 10:16:53 VCS INFO V-16-1-10493 Evaluating sirius as potential target node for group dbhost-app_sg
2013/07/05 10:16:53 VCS NOTICE V-16-1-10233 Clearing Restart attribute for group dbhost-app_sg on all nodes
2013/07/05 10:20:16 VCS ERROR V-16-1-10205 Group dbhost-app_sg is faulted on system sirius
2013/07/05 10:20:16 VCS NOTICE V-16-1-10446 Group dbhost-app_sg is offline on system sirius
2013/07/05 10:20:16 VCS INFO V-16-1-10493 Evaluating sirius as potential target node for group dbhost-app_sg
2013/07/05 10:20:16 VCS INFO V-16-1-50010 Group dbhost-app_sg is online or faulted on system sirius
2013/07/05 10:20:16 VCS INFO V-16-1-10493 Evaluating arcturus as potential target node for group dbhost-app_sg
2013/07/05 10:20:16 VCS INFO V-16-1-50010 Group dbhost-app_sg is online or faulted on system arcturus
 

 

Only zfs pools and IP prevent to go online of dbhost-app_sg on Node B:

..

2013/07/05 10:17:44 VCS WARNING V-16-10001-20002 (sirius) Zpool:zpool_limsdb-admin:online:zpool import limsdb-admin failed. Try again using the force import -f option
2013/07/05 10:17:45 VCS WARNING V-16-10001-20002 (sirius) Zpool:zpool_limsdb-archivedata:online:zpool import limsdb-archivedata failed. Try again using the force import -f option
2013/07/05 10:17:48 VCS WARNING V-16-10001-20002 (sirius) Zpool:zpool_limsdb-datafiles:online:zpool import limsdb-datafiles failed. Try again using the force import -f option
2013/07/05 10:17:54 VCS WARNING V-16-10001-20002 (sirius) Zpool:zpool_limsdb-indexfiles:online:zpool import limsdb-indexfiles failed. Try again using the force import -f option
..

013/07/05 10:16:53 VCS ERROR V-16-10001-5013 (sirius) IPMultiNICB:dbhost_ipmultinicb_VLAN10:online:This IP address is configured elsewhere. Will not online
2013/07/05 10:17:53 VCS ERROR V-16-10001-5013 (sirius) IPMultiNICB:dbhost_ipmultinicb_VLAN10:online:This IP address is configured elsewhere. Will not online

..

 

main.cf:

-------------------------------------------------------

group dbhost-app_sg (
    SystemList = { sirius = 1, arcturus = 0 }
    ContainerInfo @sirius = { Name = nssitdb01-zone, Type = Zone, Enabled = 1 }
    ContainerInfo @arcturus = { Name = nssitdb01-zone, Type = Zone, Enabled = 1 }
    AutoStartList = { arcturus, sirius }
    Administrators = { z_nssitdb01-zone_arcturus, z_nssitdb01-zone_sirius }
    )
...

requires group nssitdb01-zone_sg online local firm
--------------------------------------------------------

group nssitdb01-zone_sg (
        SystemList = { arcturus = 0, sirius = 1 }
    ContainerInfo @arcturus = { Name = nssitdb01-zone, Type = Zone, Enabled = 1 }
    ContainerInfo @sirius = { Name = nssitdb01-zone, Type = Zone, Enabled = 1 }
    Parallel = 1
    AutoStartList = { arcturus, sirius }
    Administrators = { z_nssitdb01-zone_sirius, z_nssitdb01-zone_arcturus }
    )

        FileNone nssitdb01-zone-root_FileNone (
        PathName = "/export/home/nssitdb01-zone/root/.vcs-FileNone-agent"
        )

        Zone nssitdb01-zone (
        Critical = 0
        DetachZonePath = 0
        )

        nssitdb01-zone requires nssitdb01-zone-root_FileNone
 

---------------------------------------------------------------------------------

Network setup for VMware and VCS

$
0
0
I need a solution

Hi,

I am trying to set up a cluster consisting of four VMware VMs. I am still having problems finding how to set up the network connections for these systems.
I have read that we need vNetwork Distributed Switches for this. Is that correct? We are installing this into an existing ESX infrastructure and we do not have any spare NICs anymore. So I suppose we have to move some of the NICs to the new VDS.

I would be really grateful if you could help me, or poiunt me in the right direction.

Regards
Thomas

 

ESX 5.x
VCS 6.0.2
RHEL 6

1373049398

Solaris 9 branded zones with VCS 6.0.1

$
0
0
I need a solution

After going through the upgrade process from SFHA 5.1 to 6.0.1, I've come to find out that 6.0.1 dropped support for branded zones. What is the fix for this? I have 9 different Solaris 9 branded zones that cannot be upgraded because of legacy applications. These zones all worked with no problem with SFHA 5.1. I'm still able to manually start the zones outside of the cluster, but that defeats the purpose of having a cluster.

I don't want to have to downgrade back to SFHA 5.1 if I can help it, but that's the only option I see at this point.

The automated preupgrade checks passed with no mention of dropping support for these zones. The zones were running at the time.

 

Faulted on node2

$
0
0
I need a solution

Hi

 

Faulted on node2 This message comes when i turn off the node1  to see the failover work or not  my services group and the resource not failover to node 2

 (Service Groups state is online and the system list for the  Service Groups on both nodes)

Please help .

 

Regards

How does VCS check IF different heartbeat NICs are in same network?

$
0
0
I need a solution

Cluster-A syslog shows:

Jun 23 07:22:11 IMDB1 vmunix: LLT WARNING V-14-1-10498 recvarpreq cross links? links 1 and 0 of node 1 connected to same network?
Jun 23 07:27:11 IMDB1 vmunix: LLT WARNING V-14-1-10498 recvarpreq cross links? links 1 and 0 of node 1 connected to same network?
Jun 23 07:32:11 IMDB1 vmunix: LLT WARNING V-14-1-10498 recvarpreq cross links? links 1 and 0 of node 1 connected to same network?
Jun 23 07:57:11 IMDB1 vmunix: LLT WARNING V-14-1-10498 recvarpreq cross links? links 1 and 0 of node 1 connected to same network?
Jun 23 08:12:11 IMDB1 vmunix: LLT WARNING V-14-1-10498 recvarpreq cross links? links 1 and 0 of node 1 connected to same network?
Jun 23 08:17:11 IMDB1 vmunix: LLT WARNING V-14-1-10498 recvarpreq cross links? links 0 and 1 of node 1 connected to same network?

----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

The heartbeats NICs are lan901 and lan902.

I know VCS heatbeats are broadcast packets. and LLT on each system in the cluster sends heartbeat packets out on all configured LLT interfaces every half second.

My questions are:

1) How, by what mechanism, does VCS check IF different heartbeat NICs are in same network (VLAN)?

   To which kind of address (MAC or OS device names) does the check signals send?

2) How often does it check if NICs are in same VLAN?

3) We have other systems Cluster-B and Cluster-C, they have almost the same network configurations (To be verified), but their syslogs do not have these kind of warnings.Why?

Thanks in advance!

1373419992

Service Group for Oracle Database inside the solaris zone(container)

$
0
0
I need a solution
SMON: enabling tx recovery
Database Characterset is WE8MSWIN1252
No Resource Manager plan active
replication_dependency_tracking turned off (no async multimaster replication found)
Starting background process QMNC
Thu Jul 11 14:02:03 2013
QMNC started with pid=56, OS id=13152
Completed: ALTER DATABASE OPEN
Thu Jul 11 14:02:04 2013
Starting background process CJQ0
Thu Jul 11 14:02:04 2013
CJQ0 started with pid=54, OS id=13176
Thu Jul 11 14:02:19 2013
Starting background process SMCO
Thu Jul 11 14:02:19 2013
SMCO started with pid=65, OS id=13276
Thu Jul 11 14:04:06 2013
Shutting down instance (abort)
License high water mark = 32
USER (ospid: 13498): terminating the instance
Instance terminated by USER, pid = 13498
Thu Jul 11 14:04:08 2013
Instance shutdown complete
 

Shared robotic library configuration

$
0
0
I need a solution

Hi all,

We need to install VCS 6.0 on RHEL 6.x. Both nodes has a robotic library attached. Backup software guide (Networker) talks about a special device file for it provided by cluster vendor, to use the robot as a HA resource. Has someone the procedure to configure that device? I cant found it yet.

Regards

Claudio

IMF VCS 6

$
0
0
I need a solution

i read in the release note that IMF is turned on by default. but it doesn't seem to be turned on if i didn't use the installer to configure the cluster.

i just want to verify this with people out there.

also how do you find AMF so far?

VERITAS STORAGE EXEC 5.5 REPLICATIONS PROBLEMS

$
0
0
I need a solution

Hi,

I have two nodes in cluster and quotas defined in Node A are not replicated to a Node B, when I switch the cluster resources.

Where policies and quotas are defined ?

Local:

C: \ ProgramFiles \ Veritas \ StorageExec \ DB
SCAudit.mdb or SCTrend.mdb

Should not be in a cluster resource?

 

Thank you all for your answers

 

Veritas Global Cluster Solution VS ESXi VMware with SRM

$
0
0
I need a solution

 

Dear friends.

One of our client has a plan to go on virtulization.

  •  To my knowledge two ESXi hosts with a SAN attached can give local failover of VM's from one ESXi host to another ESXi host. If we add VMware SRM(Site Recovery Manager) product between these two ESXi hosts(PRIMARY) and one ESXi host (DR) , we can establish a Global Cluster(In Symantec term) and through above solution all VM's which exist at PRIMARY site can failover to DR site if PRIMARY site goes down.
  • For establishing this environment we need the licenses of three ESXi hosts, two Vcenters and one SRM.
  • This solution also has a feature like fire drill in SFHA.
  • This solution provide failover on ESXi host failure, OS failure or any kind of hardware component failure. It does not have a super feature which can detect application failure. For this Symantec has a Application HA. The Application HA integrates with the above solution and provide failover on application failure as well.

So to my knowledge in the above virtulization scenerio we can only eligible to propose Application HA. As the above solution can provide best response for Hardware/OS failure.

 

Correct me if I am wrong for any point mentioned above. Suggestion required if we can improve the above solution with the help of Symantec product(s).

 

 

 

 

vcs storage migration

$
0
0
I need a solution

Is it possible to migrate data using vxevac on a three node vcs cluster without taking cluster down.

I have two dg's on each node active.

 

 

VCS Error Codes for all platforms

$
0
0
I need a solution

Hello Gents,

Do we have a list of all erro codes for VCS ?

also if all the error codes are generic and are common for all platforms (including Linux,Solaris, Windows,AIX)

Need this confirmation urgently, planning to design a common monitoring agent.

Best Regards,

Nimish

Removing missing drives and volumes

$
0
0
I need a solution

We have LUNs that were removed and now I can't remove the volumes or diskgroups.  I need to remove the volumes, disk groups and drives drives from Veritas Enterprise Administrator


Group Dependencies

$
0
0
I need a solution

Hi Experts,

I wanted to configure the SG dependencies with below requirement:

Application and Database should not run on same node. Please let me know the possibilities. Thanks

 

Thank You

Licenses

$
0
0
I need a solution

Hi

 

I installed VRTS_SF_HA_Solutions_6.0.1_RHEL (  Product Type Standard ) on my servers and is working good the services group and Failover everything good so is this a free version or my services will be stop after days and  need to buy the  version.

 

 

Regards

How to integrate Active Directory users to an already configured VCS two node (non-secure) cluster

$
0
0
I need a solution

Hello colleagues,

 

I have the following situation:

 

An already configured two node VCS cluster (6.01 version) in Solaris 10 Sparc. We have installed the cluster in non-secure mode. The cluster has been working and the only configured used is the "admin" built in account.

We have a Active Directory environment and we want to integrate its users to the VCS cluster. Is it possible to do that? We have to reconfigure the cluster in secure mode to do that? If yes, how to reconfigure an already configured cluster?.

 

Thanks in advance and best regards

 

Osvaldo Olmedo

Root cause needed for failed failover test

$
0
0
I need a solution

we tried to do failover , however it didnot worked

please find below logs and please help in finding the cause

 

2013/07/22 20:11:17 VCS INFO V-16-1-50859 Attempting to switch group Oss from system dukosgbs to system dukosgas
2013/07/22 20:11:17 VCS INFO V-16-1-50135 User root fired command: hagrp -switch Oss  dukosgas  from localhost
2013/07/22 20:11:17 VCS NOTICE V-16-1-50929 Initial tests indicate group Oss is able to switch to system dukosgas.  Initiating offline of group on system dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10167 Initiating manual offline of group Oss on system dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource activemq (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource alex (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource apache (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource cron (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource ddc (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource glassfish (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource imgr_httpd (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource imgr_tomcat (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource ldap_mon (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource log_service (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource netmgt_nettl (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource netmgt_ov (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource ovtrc (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource restart_mc (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource syb_log_mon (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource syb_proc_mon (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource time_service (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource trapdist (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource vrsnt_log_mon (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:45:26 VCS INFO V-16-1-50135 User root fired command: hagrp -switch Oss  dukosgas  from localhost
2013/07/22 22:19:35 VCS INFO V-16-2-13075 (dukosgbs) Resource(activemq) has reported unexpected OFFLINE 1 times, which is still within the ToleranceLimit(2).
2013/07/22 22:20:35 VCS INFO V-16-2-13075 (dukosgbs) Resource(activemq) has reported unexpected OFFLINE 2 times, which is still within the ToleranceLimit(2).
2013/07/22 22:21:35 VCS ERROR V-16-2-13067 (dukosgbs) Agent is calling clean for resource(activemq) because the resource became OFFLINE unexpectedly, on its own.
2013/07/22 22:21:36 VCS INFO V-16-2-13068 (dukosgbs) Resource(activemq) - clean completed successfully.
2013/07/22 22:21:36 VCS ERROR V-16-2-13073 (dukosgbs) Resource(activemq) became OFFLINE unexpectedly on its own. Agent is restarting (attempt number 1 of 3) the resource.
2013/07/22 22:21:36 VCS INFO V-16-10001-3 (dukosgbs) Application:activemq:online:Executed /ericsson/hacs/scripts/svc.sh
2013/07/22 22:21:37 VCS INFO V-16-2-13001 (dukosgbs) Resource(activemq): Output of the completed operation (online)
svcadm: Instance "svc:/ericsson/eric_3pp/activemq:default" is not in a maintenance or degraded state.
2013/07/22 22:21:38 VCS NOTICE V-16-2-13076 (dukosgbs) Agent has successfully restarted resource(activemq).
2013/07/22 22:23:05 VCS INFO V-16-1-50135 User root fired command: hagrp -clear Oss  dukosgbs  from localhost
2013/07/22 22:27:08 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Oss  dukosgbs  from localhost
2013/07/22 22:29:40 VCS INFO V-16-1-50135 User root fired command: hagrp -clearadminwait Oss  dukosgbs  from localhost
2013/07/22 22:37:21 VCS INFO V-16-1-50135 User root fired command: hagrp -flush ClusterService  dukosgbs  from localhost
2013/07/22 22:37:21 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Oss  dukosgbs  from localhost
2013/07/22 22:37:21 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Ossfs  dukosgbs  from localhost
2013/07/22 22:37:21 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Sybase1  dukosgbs  from localhost
2013/07/22 22:38:14 VCS INFO V-16-1-50135 User root fired command: hagrp -flush ClusterService  dukosgbs  from localhost
2013/07/22 22:38:14 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Oss  dukosgbs  from localhost
2013/07/22 22:38:14 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Ossfs  dukosgbs  from localhost
2013/07/22 22:38:14 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Sybase1  dukosgbs  from localhost
2013/07/22 22:39:08 VCS INFO V-16-1-50135 User root fired command: hares -refreshinfo activemq  from localhost
2013/07/22 22:40:06 VCS INFO V-16-1-50135 User root fired command: hares -refreshinfo activemq  localclus  from localhost
2013/07/22 22:42:09 VCS INFO V-16-1-50135 User root fired command: hares -flushinfo activemq  localclus  from localhost
2013/07/22 22:47:41 VCS INFO V-16-1-50135 User root fired command: hagrp -switch Oss  dukosgbs  from localhost
2013/07/22 22:50:52 VCS INFO V-16-1-50859 Attempting to switch group Oss from system dukosgbs to system dukosgas
2013/07/22 22:50:52 VCS INFO V-16-1-50135 User root fired command: hagrp -switch Oss  dukosgas  from localhost
2013/07/22 22:50:52 VCS NOTICE V-16-1-50929 Initial tests indicate group Oss is able to switch to system dukosgas.  Initiating offline of group on system dukosgbs

switching of service group didnot work vcs 5.0

$
0
0
I need a solution

we tried to do failover , however it didnot worked

please find below logs and please help in finding the cause

 

2013/07/22 20:11:17 VCS INFO V-16-1-50859 Attempting to switch group Oss from system dukosgbs to system dukosgas
2013/07/22 20:11:17 VCS INFO V-16-1-50135 User root fired command: hagrp -switch Oss  dukosgas  from localhost
2013/07/22 20:11:17 VCS NOTICE V-16-1-50929 Initial tests indicate group Oss is able to switch to system dukosgas.  Initiating offline of group on system dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10167 Initiating manual offline of group Oss on system dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource activemq (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource alex (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource apache (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource cron (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource ddc (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource glassfish (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource imgr_httpd (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource imgr_tomcat (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource ldap_mon (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource log_service (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource netmgt_nettl (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource netmgt_ov (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource ovtrc (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource restart_mc (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource syb_log_mon (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource syb_proc_mon (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource time_service (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource trapdist (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:11:17 VCS NOTICE V-16-1-10300 Initiating Offline of Resource vrsnt_log_mon (Owner: unknown, Group: Oss) on System dukosgbs
2013/07/22 20:45:26 VCS INFO V-16-1-50135 User root fired command: hagrp -switch Oss  dukosgas  from localhost
2013/07/22 22:19:35 VCS INFO V-16-2-13075 (dukosgbs) Resource(activemq) has reported unexpected OFFLINE 1 times, which is still within the ToleranceLimit(2).
2013/07/22 22:20:35 VCS INFO V-16-2-13075 (dukosgbs) Resource(activemq) has reported unexpected OFFLINE 2 times, which is still within the ToleranceLimit(2).
2013/07/22 22:21:35 VCS ERROR V-16-2-13067 (dukosgbs) Agent is calling clean for resource(activemq) because the resource became OFFLINE unexpectedly, on its own.
2013/07/22 22:21:36 VCS INFO V-16-2-13068 (dukosgbs) Resource(activemq) - clean completed successfully.
2013/07/22 22:21:36 VCS ERROR V-16-2-13073 (dukosgbs) Resource(activemq) became OFFLINE unexpectedly on its own. Agent is restarting (attempt number 1 of 3) the resource.
2013/07/22 22:21:36 VCS INFO V-16-10001-3 (dukosgbs) Application:activemq:online:Executed /ericsson/hacs/scripts/svc.sh
2013/07/22 22:21:37 VCS INFO V-16-2-13001 (dukosgbs) Resource(activemq): Output of the completed operation (online)
svcadm: Instance "svc:/ericsson/eric_3pp/activemq:default" is not in a maintenance or degraded state.
2013/07/22 22:21:38 VCS NOTICE V-16-2-13076 (dukosgbs) Agent has successfully restarted resource(activemq).
2013/07/22 22:23:05 VCS INFO V-16-1-50135 User root fired command: hagrp -clear Oss  dukosgbs  from localhost
2013/07/22 22:27:08 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Oss  dukosgbs  from localhost
2013/07/22 22:29:40 VCS INFO V-16-1-50135 User root fired command: hagrp -clearadminwait Oss  dukosgbs  from localhost
2013/07/22 22:37:21 VCS INFO V-16-1-50135 User root fired command: hagrp -flush ClusterService  dukosgbs  from localhost
2013/07/22 22:37:21 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Oss  dukosgbs  from localhost
2013/07/22 22:37:21 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Ossfs  dukosgbs  from localhost
2013/07/22 22:37:21 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Sybase1  dukosgbs  from localhost
2013/07/22 22:38:14 VCS INFO V-16-1-50135 User root fired command: hagrp -flush ClusterService  dukosgbs  from localhost
2013/07/22 22:38:14 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Oss  dukosgbs  from localhost
2013/07/22 22:38:14 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Ossfs  dukosgbs  from localhost
2013/07/22 22:38:14 VCS INFO V-16-1-50135 User root fired command: hagrp -flush Sybase1  dukosgbs  from localhost
2013/07/22 22:39:08 VCS INFO V-16-1-50135 User root fired command: hares -refreshinfo activemq  from localhost
2013/07/22 22:40:06 VCS INFO V-16-1-50135 User root fired command: hares -refreshinfo activemq  localclus  from localhost
2013/07/22 22:42:09 VCS INFO V-16-1-50135 User root fired command: hares -flushinfo activemq  localclus  from localhost
2013/07/22 22:47:41 VCS INFO V-16-1-50135 User root fired command: hagrp -switch Oss  dukosgbs  from localhost
2013/07/22 22:50:52 VCS INFO V-16-1-50859 Attempting to switch group Oss from system dukosgbs to system dukosgas
2013/07/22 22:50:52 VCS INFO V-16-1-50135 User root fired command: hagrp -switch Oss  dukosgas  from localhost
2013/07/22 22:50:52 VCS NOTICE V-16-1-50929 Initial tests indicate group Oss is able to switch to system dukosgas.  Initiating offline of group on system dukosgbs

 

 

=====================================================================

 

the resources ware initiating to offline. But they did not come down.
 

 

root@dukosgbs> hastatus -sum
 
 
 
 -- SYSTEM STATE
 
 -- System               State                Frozen
 
 
 
 A  dukosgbs             RUNNING              0
 
 
 
 -- GROUP STATE
 
 -- Group           System               Probed     AutoDisabled    State
 
 
 
 B  ClusterService  dukosgbs             Y          N               ONLINE
 
 B  Oss             dukosgbs             Y          N               ONLINE|STOPPING
 
 B  Ossfs           dukosgbs             Y          N               ONLINE
 
 B  Sybase1         dukosgbs             Y          N               ONLINE
 
 
 
 -- RESOURCES OFFLINING
 
 -- Group           Type            Resource             System               IState
 
 
 
 F  Oss             Application     activemq             dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     alex                 dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     apache               dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     cron                 dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     ddc                  dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     glassfish            dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     imgr_httpd           dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     imgr_tomcat          dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     ldap_mon             dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     log_service          dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     netmgt_nettl         dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     netmgt_ov            dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     ovtrc                dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     restart_mc           dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     syb_log_mon          dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     syb_proc_mon         dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     time_service         dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     trapdist             dukosgbs             W_OFFLINE_PROPAGATE
 
 F  Oss             Application     vrsnt_log_mon        dukosgbs             W_OFFLINE_PROPAGATE
 
 
 
 -- WAN HEARTBEAT STATE
 
 -- Heartbeat       To                   State
 
 
 
 L  Icmp            gran_cluster1        ALIVE
 
 
 
 -- REMOTE CLUSTER STATE
 
 -- Cluster         State
 
 
 
 M  gran_cluster1   RUNNING
 
 
 
 -- REMOTE SYSTEM STATE
 
 -- cluster:system       State                Frozen
 
 
 
 N  gran_cluster1:dukosgas RUNNING              0

1375291747
Viewing all 543 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>