Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Author
Team
Date of Creation
Reviewed by
Email-id
K.Kandasamy
TIS - Unix Team
13th December 2007
Ramachandra Pargaonkar
Kandasamy.kumaravel@wipro.com
Note: For RAC environments, the cluster name (similar to the hostid) gets stamped onto the
provide region of the disks. Therefore, if you change the name of the cluster also update the
cluster name on the disks by following these steps:
1. Confirm you are on the Master Node:
# vxdctl -c mode
2. Update the cluster name stamped on the disks:
# vxdg deport [disk_group]
# vxdg -Cs import [disk_group]
===============================================================
2. VCS CRITICAL V-16-1-10029 VxFEN driver not configured. VCS Stopping.
Manually restart VCS after configuring fencing
After rebooting the nodes in a cluster, Veritas Cluster Server (VCS) fails to start and the
following messages are seen in the /var/VRTSvcs/log/engine_A.log file:
2007/10/22 05:16:47 VCS CRITICAL V-16-1-10037 VxFEN driver not configured.
Retrying...
2007/10/22 05:17:02 VCS CRITICAL V-16-1-10037 VxFEN driver not configured.
Retrying...
2007/10/22 05:17:17 VCS CRITICAL V-16-1-10037 VxFEN driver not configured.
Retrying...
2007/10/22 05:17:33 VCS CRITICAL V-16-1-10037 VxFEN driver not configured.
Retrying...
2007/10/22 05:17:48 VCS CRITICAL V-16-1-10037 VxFEN driver not configured.
Retrying...
2007/10/22 05:18:03 VCS CRITICAL V-16-1-10037 VxFEN driver not configured.
Retrying...
2007/10/22 05:18:18 VCS CRITICAL V-16-1-10029 VxFEN driver not configured. VCS
Stopping. Manually restart VCS after configuring fencing
Attempting to start I/O Fencing manually results in the following error:
# ./S97vxfen start
Details:
The issue:
An error message in /var/VRTSvcs/log/engine_A.log states that the mount option is
incompatible with the file system.
Change:
The largefiles option was added to the MountOpt attribute for the mount resource.
Resolution:
The largefiles option must be enabled for the file system at the operating system level before
it can be configured for largefiles within Veritas Cluster Server (VCS).
Enable largefiles with this command:
/usr/lib/fs/vxfs/fsadm -o largefiles <mount>
14) Kernel message: Dazed and confused, but trying to continue
Details:
Symptoms:
System panic with error messages on boot:
kernel: LLT INFO V-14-1-10009 LLT Protocol available
kernel: device eth1 entered promiscuous mode
kernel: tg3: eth1: Link is up at 1000 Mbps, full duplex.
kernel: tg3: eth1: Flow control is on for TX and on for RX.
kernel: device eth1 left promiscuous mode
kernel: Uhhuh. NMI received for unknown reason 35 on CPU 0.
kernel: Dazed and confused, but trying to continue
kernel: Do you have a strange power saving mode enabled?
Cause:
Defective hardware was used in the loading of the LLT module.
Resolution:
Replace CPU 0.
===============================================================
15) Multi Network Interface Card B (MultiNICB) resource demands high CPU time
with Solaris IP Multipathing (IPMP)
Details
Solaris IP Multipathing (IPMP) is in use and the MultiNICB resource is configured with the
UseMpathd attribute enabled (set to 1).
Cause
During every monitor cycle, the MultiNICB agent checks the system process table for the
IPMP daemon process in mpathd.
If several MultiNICB resources are configured in the cluster, the agent checks for the IPMP
daemon many times every minute, resulting in a higher CPU demand.
Workaround
To decrease CPU demand:
Increase the MonitorInterval attribute for the MultiNICB resource from the default 10 seconds
to 30 seconds.
In cluster configurations that have more than three MultiNICB resources, change the
NumThreads attribute from the default of 10 to 1 or 2.
Enhancement request e426856 addresses this issue in a future product or patch release
4.1MP4+e1023246
2007-May-14
VCS RELEASE:
4.1MP4
LINUX RELEASE:
SLES 10 SP1
RELEVANT ARCHITECTURES:
ETRACK REFERENCE:
1023246
PROBLEM DESCRIPTION:
'Stale NFS handle' errors seen on accessing files exported by NFS clients
when a service group configured with NFS fails over
PATCH CONTAINS
-------------.
|__ online
|__ README
PATCH INSTALLATION INSTRUCTIONS:
-------------------------------Installed this patch after installing Veritas Cluster Server 4.1 MP4, following these steps:
The Default value of $VCS_HOME is /opt/VRTSvcs
1. Log in as superuser to the system where the point patch is to be installed.
2. Go to the directory $VCS_HOME/bin/NFS:
#cd $VCS_HOME/bin/NFS
3. Copy online as online.orig on all nodes of the cluster:
#cp online online.orig
4. On each node of the cluster, copy the "online" from this patch to
$VCS_HOME/bin/NFS/online:
#cp /PointPatchDir/online ./online
===============================================================
17) Veritas Cluster Server (VCS) I/O Fencing parameters for racing (Solaris)
Details:
When communication between cluster nodes fails, causing the cluster to be divided into subclusters, these sub-clusters start a race to grab coordinator disks for data protection (VCS I/O
Fencing). vxfen has a mechanism that enables cluster administrators to give larger subclusters better odds to win this race. This document describes the differences in
implementation between VCS versions and their tunable parameters.
Note: While this mechanism can be used to give larger sub-clusters much better odds to win
the race condition, it cannot be used to guarantee that the larger sub-cluster will always win.
1. How to give the odds
Prior to 4.1 MP2
If the number of nodes in a sub-cluster is less than the number of nodes leaving the original
cluster, the sub-cluster repeats reading the coordinator disks to delay the start of the race. By
default, the number of reads is calculated as cube of (the number of leaving nodes). For
example, if a 5-node cluster is divided into a 3-node and a 2-node cluster, the 2-node subcluster repeats reading coordinator disks 27 (= 3 cubed) times. A tunable parameter,
max_read_coord_disk, can be used to change this value, as described later.
4.1 MP2 and 5.0 or later