Hi,
Lanman services from VCS and lanmanserver from windows are same?
Hi,
Lanman services from VCS and lanmanserver from windows are same?
Hi,
We receive VCS CRITICAL CPU usage errors very often:
Jan 8 01:08:45 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 01:08:45 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 01:08:52 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 01:08:52 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 01:08:57 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 01:08:57 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 01:30:43 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 01:30:43 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 01:51:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS ERROR V-16-1-1
3027 (LHDBEAP01) Resource(app) - monitor procedure did not complete within the expected time.
Jan 8 06:10:01 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 06:10:01 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 08:17:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 91%
Jan 8 08:24:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 08:34:02 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 08:45:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 93%
Jan 8 08:47:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 96%
Jan 8 09:00:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 94%
Jan 8 09:01:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 96%
Jan 8 09:05:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 99%
Jan 8 09:26:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 96%
Jan 8 09:28:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 95%
Jan 8 09:40:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 91%
Jan 8 09:54:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 93%
Jan 8 09:55:02 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 94%
Jan 8 09:58:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 10:09:02 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 99%
Jan 8 10:12:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 91%
Jan 8 10:17:30 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS ERROR V-16-1-1
3027 (LHDBEAP01) Resource(app) - monitor procedure did not complete within the expected time.
Jan 8 10:23:31 LHDBEDB01 last message repeated 2 times
Jan 8 10:24:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 95%
Jan 8 10:28:12 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 10:28:12 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 10:28:15 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 10:28:15 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 10:28:18 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 10:28:18 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 10:28:49 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 10:28:49 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 10:30:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 92%
Jan 8 10:53:02 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 98%
Jan 8 10:54:02 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 96%
Jan 8 10:59:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 91%
Jan 8 11:22:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 95%
Jan 8 11:44:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 99%
Jan 8 11:49:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 96%
Jan 8 11:56:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 93%
Jan 8 12:25:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 94%
Jan 8 12:26:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 91%
Jan 8 12:41:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 95%
Jan 8 12:47:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 94%
Jan 8 12:49:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 99%
Jan 8 13:02:02 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 94%
Jan 8 13:04:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 92%
Jan 8 13:12:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 91%
Jan 8 13:16:02 LHDBEDB01 last message repeated 2 times
Jan 8 13:36:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 97%
Jan 8 13:42:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 99%
Jan 8 13:45:32 LHDBEDB01 last message repeated 1 time
Jan 8 13:47:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 13:51:02 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 99%
Jan 8 14:02:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 96%
Jan 8 14:03:47 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 14:03:47 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 14:07:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 14:11:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 99%
Jan 8 14:20:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 97%
Jan 8 14:24:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 14:25:31 LHDBEDB01 last message repeated 1 time
Jan 8 14:32:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 92%
Jan 8 14:35:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 97%
Jan 8 14:42:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS ERROR V-16-1-1
3027 (LHDBEAP01) Resource(app) - monitor procedure did not complete within the expected time.
Jan 8 14:45:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 98%
Jan 8 14:47:02 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 92%
Jan 8 14:48:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 96%
Jan 8 14:51:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 97%
Jan 8 14:53:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 15:02:01 LHDBEDB01 last message repeated 2 times
Jan 8 15:10:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 15:12:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 91%
Jan 8 15:13:33 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS ERROR V-16-1-1
3027 (LHDBEAP01) Resource(app) - monitor procedure did not complete within the expected time.
Jan 8 15:20:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 93%
Jan 8 15:22:32 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 97%
Jan 8 15:25:02 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 15:29:01 LHDBEDB01 last message repeated 4 times
Jan 8 15:33:31 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 15:38:31 LHDBEDB01 last message repeated 2 times
Jan 8 15:41:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 96%
Jan 8 15:44:02 LHDBEDB01 last message repeated 1 time
Jan 8 15:50:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 92%
Jan 8 15:51:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 100%
Jan 8 15:51:24 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 15:51:24 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 15:51:30 LHDBEDB01 llt: [ID 140958 kern.notice] LLT INFO V-14-1-10205 link 2 (ce0) node 1 in trouble
Jan 8 15:51:30 LHDBEDB01 llt: [ID 860062 kern.notice] LLT INFO V-14-1-10024 link 2 (ce0) node 1 active
Jan 8 15:57:01 LHDBEDB01 Had[377]: [ID 702911 daemon.notice] VCS CRITICAL V-16-1-50086 CPU usage on LHDBEDB01 is 99%
bash-2.05$
root@LHDBEAP01 # pkginfo -l NTAPnfsa
PKGINST: NTAPnfsa
NAME: NetApp NFS Client Agent for Veritas Cluster Server by Symantec
CATEGORY: optional
ARCH: sun4u
VERSION: 5.0.0
BASEDIR: /
VENDOR: Network Appliance, Inc.
DESC: NetApp NFS Client Agent for Veritas Cluster Server by Symantec
PSTAMP: NTAPnfsa:693884:20070816.1459
INSTDATE: Dec 18 2013 03:54
STATUS: completely installed
FILES: 26 installed pathnames
7 shared pathnames
10 directories
2 executables
242 blocks used (approx)
bash-2.05$ sar 1 10
SunOS LHDBEDB01 5.9 Generic_117171-17 sun4u 01/08/2015
16:11:27 %usr %sys %wio %idle
16:11:28 71 2 0 27
16:11:29 56 5 0 39
16:11:30 51 4 0 45
16:11:31 78 9 0 13
16:11:32 52 0 0 48
16:11:33 51 2 0 47
16:11:34 56 5 0 39
16:11:35 50 1 0 49
16:11:36 58 6 0 36
16:11:37 50 1 0 49
Average 57 3 0 39
bash-2.05$
Hi
I have a service group named redsg running which is online on node1 and and offline on node 2.
I froze resg and ran hastop -all -force ,and started it back again on both nodes
the service group state is online now on node1 and offline on node2 as earlier
But when i reboot the node the service group goes to offline on both nodes.
Any idea what makes this difference.
Thanks
dilip
Hi,
i am trying to configure the Oracle Apps to be under cluster when i put all the required vairables required by the resource the cluster couldn't be able detect that the processes is online ,here under the variables:
Oracle Apps Version | 12 |
User Name | applprod |
Oracle Home | /binaries/apps/tech_st/10.1.2 |
Script Home | $INST_TOP/admin/scripts SCRIPT NAME: adstrtal.sh |
Server Type | Webserver |
Monitor | /export/home/applprod/monitor/orapps.prob |
Env | $INST_TOP/ora/10.1.2/ERPPROD_erpapp-lh.env |
Any Idea about what shoud i do.
Regards,
I have a three-node cluster configuration: vcs1, vcs2, vcs3. Using three ISCSI disks (by SCST iscsi target simulator) as fendg.
# vxfenadm -d
I/O Fencing Cluster Information:
================================
Fencing Protocol Version: 201
Fencing Mode: SCSI3
Fencing SCSI3 Disk Policy: dmp
Cluster Members:
* 0 (vcs1)
1 (vcs2)
2 (vcs3)
RFSM State Information:
node 0 in state 8 (running)
node 1 in state 8 (running)
node 2 in state 8 (running)
# vxfenadm -s all -f /etc/vxfentab
Device Name: /dev/vx/rdmp/disk_1s3
Total Number Of Keys: 3
key[0]:
[Numeric Format]: 86,70,49,48,53,50,48,48
[Character Format]: VF105200
* [Node Format]: Cluster ID: 4178 Node ID: 0 Node Name: vcs1
key[1]:
[Numeric Format]: 86,70,49,48,53,50,48,49
[Character Format]: VF105201
* [Node Format]: Cluster ID: 4178 Node ID: 1 Node Name: vcs2
key[2]:
[Numeric Format]: 86,70,49,48,53,50,48,50
[Character Format]: VF105202
* [Node Format]: Cluster ID: 4178 Node ID: 2 Node Name: vcs3
Device Name: /dev/vx/rdmp/disk_0s3
Total Number Of Keys: 3
key[0]:
[Numeric Format]: 86,70,49,48,53,50,48,48
[Character Format]: VF105200
* [Node Format]: Cluster ID: 4178 Node ID: 0 Node Name: vcs1
key[1]:
[Numeric Format]: 86,70,49,48,53,50,48,49
[Character Format]: VF105201
* [Node Format]: Cluster ID: 4178 Node ID: 1 Node Name: vcs2
key[2]:
[Numeric Format]: 86,70,49,48,53,50,48,50
[Character Format]: VF105202
* [Node Format]: Cluster ID: 4178 Node ID: 2 Node Name: vcs3
Device Name: /dev/vx/rdmp/disk_2s3
Total Number Of Keys: 3
key[0]:
[Numeric Format]: 86,70,49,48,53,50,48,48
[Character Format]: VF105200
* [Node Format]: Cluster ID: 4178 Node ID: 0 Node Name: vcs1
key[1]:
[Numeric Format]: 86,70,49,48,53,50,48,49
[Character Format]: VF105201
* [Node Format]: Cluster ID: 4178 Node ID: 1 Node Name: vcs2
key[2]:
[Numeric Format]: 86,70,49,48,53,50,48,50
[Character Format]: VF105202
* [Node Format]: Cluster ID: 4178 Node ID: 2 Node Name: vcs3
# lltstat -l
LLT link information:
link 0 eth1 on ether hipri
mtu 1500, sap 0xcafe, broadcast FF:FF:FF:FF:FF:FF, addrlen 6
txpkts 190145 txbytes 23138275
rxpkts 174420 rxbytes 11540391
latehb 0 badcksum 0 errors 0
link 1 eth0 on ether lowpri
mtu 1500, sap 0xcafe, broadcast FF:FF:FF:FF:FF:FF, addrlen 6
txpkts 71940 txbytes 3495901
rxpkts 73537 rxbytes 3617055
latehb 0 badcksum 0 errors 0
After I disconnect the network link of vcs3, vcs1 take over the application running on vcs3 . And after waited for serval miniutes and show log "VCS waiting for I/O fencing to be completed" , vcs3 shown kernel panic message like this:
BUG: unable to handle kernel paging request at ffffffff00000019
[32353.581223] IP: [<ffffffff810399b5>] task_rq_lock+0x35/0x90
[32353.581991] PGD 1806067 PUD 0
[32353.582446] Oops: 0000 [#1] SMP
[32353.582928] last sysfs file: /sys/devices/system/node/node0/cpumap
[32353.583751] CPU 0
[32353.584031] Modules linked in: vxodm(PN) vxfen(PN) dmpjbod(PN) dmpap(PN) dmpaa(PN) vxspec(PN) vxio(PN) vxdmp(PN) binfmt_misc snd_pcm_oss snd_mixer_oss snd_seq snd_seq_device gab(PN) crc32
c iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi af_packet llt(PN) microcode amf(PN) fuse loop vxportal(PN) fdd(PN) vxfs(PN) exportfs dm_mod virtio_console virtio_balloon virtio_net rt
c_cmos snd_hda_intel rtc_core snd_hda_codec rtc_lib snd_hwdep snd_pcm tpm_tis virtio_pci snd_timer tpm sym53c8xx virtio_ring snd button sg tpm_bios floppy pcspkr virtio scsi_transport_spi i2
c_piix4 soundcore i2c_core snd_page_alloc uhci_hcd sd_mod crc_t10dif ehci_hcd usbcore edd ext3 mbcache jbd fan processor ide_pci_generic piix ide_core ata_generic ata_piix libata scsi_mod th
ermal thermal_sys hwmon
[32353.584031] Supported: Yes, External
[32353.584031] Pid: 4730, comm: vxfen Tainted: P 2.6.32.12-0.7-default #1 Bochs
[32353.584031] RIP: 0010:[<ffffffff810399b5>] [<ffffffff810399b5>] task_rq_lock+0x35/0x90
[32353.584031] RSP: 0018:ffff88006c6b5cc0 EFLAGS: 00010086
[32353.584031] RAX: ffffffff00000001 RBX: 0000000000013680 RCX: dead000000100100
[32353.584031] RDX: 0000000000000000 RSI: ffff88006c6b5d00 RDI: ffffffff81ab2df0
[32353.584031] RBP: ffff88006c6b5ce0 R08: 00000000000005db R09: 000000000000000a
[32353.584031] R10: 0000000000000000 R11: 0000000000000000 R12: 0000000000013680
[32353.584031] R13: ffffffff81ab2df0 R14: ffff88006c6b5d00 R15: 0000000000000000
[32353.584031] FS: 0000000000000000(0000) GS:ffff880006200000(0000) knlGS:0000000000000000
[32353.584031] CS: 0010 DS: 002b ES: 002b CR0: 000000008005003b
[32353.584031] CR2: ffffffff00000019 CR3: 0000000037d1b000 CR4: 00000000000406f0
[32353.584031] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[32353.584031] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
[32353.584031] Process vxfen (pid: 4730, threadinfo ffff88006c6b4000, task ffff88006d98c580)
[32353.584031] Stack:
[32353.584031] 0000000000000000 00000001007a4653 ffffffff81ab2df0 ffff880062de34b0
[32353.584031] <0> ffff88006c6b5d30 ffffffff81040e5a 000000000000000f ffff88007c0de148
[32353.584031] <0> 0000000000000086 ffff88007c0de140 00000001007a4653 0000000000000001
[32353.584031] Call Trace:
[32353.584031] [<ffffffff81040e5a>] try_to_wake_up+0x4a/0x340
[32353.584031] [<ffffffff810682a8>] up+0x48/0x50
[32353.584031] [<ffffffffa0d5ed4a>] vxfen_bcast_lost_race_msg+0x8a/0x1b0 [vxfen]
[32353.584031] [<ffffffffa0d5f63d>] vxfen_grab_coord_pt_30+0x76d/0x830 [vxfen]
[32353.584031] [<ffffffffa0d5fbe7>] vxfen_grab_coord_pt+0x87/0x1a0 [vxfen]
[32353.584031] [<ffffffffa0d6eb7c>] vxfen_msg_node_left_ack+0x22c/0x330 [vxfen]
[32353.584031] [<ffffffffa0d70f22>] vxfen_process_client_msg+0x7d2/0xb30 [vxfen]
[32353.584031] [<ffffffffa0d716db>] vxfen_vrfsm_cback+0x45b/0x17b0 [vxfen]
[32353.584031] [<ffffffffa0d8cb20>] vrfsm_step+0x1b0/0x3b0 [vxfen]
[32353.584031] [<ffffffffa0d8ee1c>] vrfsm_recv_thread+0x32c/0x970 [vxfen]
[32353.584031] [<ffffffffa0d8f5b4>] vxplat_lx_thread_base+0xa4/0x100 [vxfen]
[32353.584031] [<ffffffff81003fba>] child_rip+0xa/0x20
[32353.584031] Code: 6c 24 10 4c 89 74 24 18 49 89 fd 48 89 1c 24 49 89 f6 4c 89 64 24 08 49 c7 c4 80 36 01 00 9c 58 fa 49 89 06 49 8b 45 08 4c 89 e3 <8b> 40 18 48 03 1c c5 40 dc 91 81 48 89
df e8 68 d5 35 00 49 8b
[32353.584031] RIP [<ffffffff810399b5>] task_rq_lock+0x35/0x90
[32353.584031] RSP <ffff88006c6b5cc0>
[32353.584031] CR2: ffffffff00000019
Hello,
In order to mange IBM HTTPServer under VCS. Do we need to use Application Agent or Apache agent? Which one would you recommend?
/opt/apps/IBM/HTTPServer
Hi,
Symantec (Veritas) Cluster Server license (to run on an HP server)
The cluster server licenses will be running on 2 virtual Windows servers each with 8 cores.
Is this enough information?
What is the difference between attributes MountPoint and MountPath of Mount Type resources?
I am able to find MountPath in Windows while on Linux I can see MountPoint only.
I have a script using a command which is trying to get MountPath but is failing on Linux.
The command is given below:
hares -display -attribute MountPath -type Mount
Hi ,
Hardware : Hp blade Gen8 .
VCS : 2 node cluster
Application Resources from one node are switched over to other node automatically and when i switched back to original node its failing .
I can see the below error messages in /var/adm/messages.
an 20 07:24:56 atrcxb3213 last message repeated 10745 times
Jan 20 07:24:56 atrcxb3213 vxvm:vxconfigd: [ID 702911 daemon.notice] V-5-1-768 Offlining config copy 1 on disk c0t0d26s2:
Jan 20 07:24:56 atrcxb3213 vxvm:vxconfigd: [ID 702911 daemon.notice] Reason: Disk write failure
Jan 20 07:24:56 atrcxb3213 vxvm:vxconfigd: [ID 702911 daemon.warning] V-5-1-579 Disk group ossdg: Errors in some configuration copies:
Jan 20 07:24:56 atrcxb3213 vxvm:vxconfigd: [ID 702911 daemon.warning] Disk c0t0d26s2, copy 1: Block 0: Disk write failure
Jan 20 07:24:56 atrcxb3213 vxvm:vxconfigd: [ID 702911 daemon.notice] V-5-1-0 Disk group import of ossdg succeeded.
Jan 20 07:24:58 atrcxb3213 scsi: [ID 107833 kern.warning] WARNING: /pci@0,0/pci8086,3c04@2/pci103c,338e@0,1/fp@0,0/disk@w5006016e46e011d3,1a (sd4):
Jan 20 07:24:58 atrcxb3213 SCSI transport failed: reason 'timeout': giving up
Jan 20 07:24:58 atrcxb3213 vxdmp: [ID 443116 kern.notice] NOTICE: VxVM vxdmp V-5-0-0 i/o error occured (errno=0x6) on dmpnode 264/0x103
Jan 20 07:24:58 atrcxb3213 last message repeated 4 times
Jan 20 07:24:58 atrcxb3213 vxio: [ID 262433 kern.warning] WARNING: VxVM vxio V-5-3-1143 voldiosiodone_fun: Hung io returned; disk = c0t0d26s2, hung
ios remaining 4
Jan 20 07:24:58 atrcxb3213 vxio: [ID 262433 kern.warning] WARNING: VxVM vxio V-5-3-1143 voldiosiodone_fun: Hung io returned; disk = c0t0d26s2, hung
ios remaining 3
Jan 20 07:24:58 atrcxb3213 vxio: [ID 262433 kern.warning] WARNING: VxVM vxio V-5-3-1143 voldiosiodone_fun: Hung io returned; disk = c0t0d26s2, hung
ios remaining 2
Jan 20 07:23:27 atrcxb3213 vxdmp: [ID 447055 kern.notice] NOTICE: VxVM vxdmp V-5-0-0 failover initiated for 264/0x100
Jan 20 07:23:27 atrcxb3213 vxdmp: [ID 447055 kern.notice] NOTICE: VxVM vxdmp V-5-0-0 curpri set to secondary for 264/0x100
Jan 20 07:23:52 atrcxb3213 AgentFramework[6534]: [ID 702911 daemon.notice] VCS ERROR V-16-2-13027 Thread(13) Resource(sybasedg) - monitor procedure
did not complete within the expected time.
Jan 20 07:23:52 atrcxb3213 Had[6465]: [ID 702911 daemon.notice] VCS ERROR V-16-2-13027 (atrcxb3213) Resource(sybasedg) - monitor procedure did not
complete within the expected time.
Jan 20 07:23:58 atrcxb3213 scsi: [ID 107833 kern.warning] WARNING: /pci@0,0/pci8086,3c04@2/pci103c,338e@0,1/fp@0,0/disk@w5006016e46e011d3,1a (sd4):
Jan 20 07:23:58 atrcxb3213 Error for Command: write(10) Error Level: Retryable
Jan 20 07:23:58 atrcxb3213 scsi: [ID 107833 kern.notice] Requested Block: 143404 Error Block: 143404
Jan 20 07:23:58 atrcxb3213 scsi: [ID 107833 kern.notice] Vendor: DGC Serial Number: 1A0000D980CL
Jan 20 07:23:58 atrcxb3213 scsi: [ID 107833 kern.notice] Sense Key: Unit_Attention
Jan 20 07:23:58 atrcxb3213 scsi: [ID 107833 kern.notice] ASC: 0x2a (asymmetric access state changed), ASCQ: 0x6, FRU: 0x0
Jan 20 07:24:42 atrcxb3213 AgentFramework[6530]: [ID 702911 daemon.notice] VCS ERROR V-16-2-13027 Thread(1888) Resource(add_swap) - monitor procedu
re did not complete within the expected time.
Regards ,
Saty
Team,
I have gone through the doc to understand the functionalities of Multinic in vcs,however unable to get it completely.
Could you please explain step by step about the Multinic and how to configure its resources?
I have to give training in multinic concept theoritically and to implement it.
Thanks..
Allaboutunix
Hello,
I have setup websphhere using Applicaiton agent on ApplicaitonHA and it is working fine. Once I se it up on 2 node cluster, once the critical resource is fualted is it not failing over. Do I need to use WebSphere as an agent? Or Application agent has capability to detect fault error for websphere application?
Hello,
on 2 servers OS Linux Red Hat 6.3, I've got a VRTS 6.0
For on Application, when I put it offline and then online on the same server (with all other resources online on this server) it start well.
But when I test a 'switch to' of all the service group on the other node it doesn't start properly (and all the other resources started )
The application is link with 3 Mount ressources and 1 IP ressource
we set attributes critical to false, and we set UseSUDash to true.
The StartProgram script is supposed to start several processes, with a offline, online action all the processes are started, with a 'switch to' action, only half of them are started.
No interesting log in the Application side.
Any suggestion to debug will be appreciated.
If we have 2 node cluster confiured and if we have 2 node names in systemList, would service group come up on node 1, If VCS down on node2, when LLT/GAB is up.
/etc/gabtab
/sbin/gabconfig -c -n2
SystemList = { node1 = 0, node2=1}
AutoStartList = {node1}
I have a requirement where I need to update the add the routes for my application to connect to the network and source systems everytime the failover or switchover happens.
If Application is running on node 1 then following routes should be added
route add -net IP netmask 255.255.255.255 10.225.230.1 gw dev bond0: 1
route add -net IP1 netmask 255.255.255.255 10.225.230.1 gw dev bond0: 1
route add -net IP2 netmask 255.255.255.255 10.225.230.1 gw dev bond0: 1
If Application is running on node 2 then following routes should be added
route add -net IP netmask 255.255.255.255 10.225.230.1 gw dev bond0: 0
route add -net IP1 netmask 255.255.255.255 10.225.230.1 gw dev bond0: 0
route add -net IP2 netmask 255.255.255.255 10.225.230.1 gw dev bond0: 0
Guys,
Could you please help me understand the need of using script ag_i18n_inc.sh loacted in $VCS_HOME/bin
Environment
SFHA version = 5.1
OS = 5.9 Solaris SPARC
Please freeze the cluster before doing this activity
Agenda
Suppose you need to replace the bge2 Ethernet with ce0 (ONLINE without restarting gab, llt and hastop -all -force)
hi,
in a vcs node i have an oracle instance running.if by accident one process of this instance is killed,how i fix this?should i freeze the service group and then stop and start the instance using the scripts?
if i shutdown a service group on an vcs node then this service group will be stared automatically on oter node?
tnx so much.
Hello,
I want your help about auditing an existing Veritas Cluster.
Can you please provide me some examples and ideas.
Thank you in advance.
hi,
i wish to force a crash dump on a vcs node.how should i perform it?
should i failover first the service group or no?
then i force the crash dump on this node.
if i do not failover the service group,then:
-if the node where i forced the crash dump will shutdown,then if i like to start the service group on other node-i should import forcefully the disk group and maybe other cleanup things for the application to start.
-if the node where i forced the crash dump reboots,and join to the vcs,then the service group will be brought up on this node automatically.
please comment.
tnx so much.
Hi All,
I am trying to install veritas storage foundation vxvm and vcs in linux rhel 6.4 OS. This is a virtual machine running on vmware. While installation veritas storage foundation (vxvm), everything got installed sucessfully but VRTSdbed & VRTSsfmh didnot installed. I have no idea what is the exact issue and resolution of this issue. can any help me on this issu
Symantec Storage Foundation Install did enot complete successfully
VRTSdbed rpm failed to install on vcsnode1
VRTSsfmh rpm failed to install on vcsnode1
rpms/patches failed to install. Do you want to exit installer? [y,n,q] (y)