Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
3 Release Notes
Description
Update eligibility
Customers running HP StoreAll OS Software 6.2.x, 6.1.x, 6.0.x, 5.6.x, 5.5.x, or 5.4.x are eligible
for the update. Refer to the administrator guide for your storage system for specific update requirements
and procedures. Customers running IBRIX OS Software 5.3.2 or earlier versions should contact HP
Support to determine compatibility before updating their software.
Supersedes
6.2.2 (build 6.2.386)
nl
Product models
HP StoreAll OS Software
Devices supported
HP StoreAll 9300 Storage Gateway
HP StoreAll 9320 Storage
HP IBRIX X9720 Storage
HP StoreAll 9730 Storage
Operating systems
StoreAll 9000 supported devices use the Red Hat Enterprise Linux 5.5 (64 bit) operating system.
Description 3
Other supported software
Linux 9000
nl
Red Hat Enterprise Linux 5.2, 5.3, 5.4, 5.5, 5.6, 5.7, 5.8, 6.1, 6.2 (all 64 bit)
clients
SUSE Linux Enterprise Server 11 (64 bit)
SUSE Linux Enterprise Server 10 SP3 (64 bit)
openSUSE 11.1 (64 bit)
CentOS 5.2, 5.3, 5.4, 5.5, 5.6, 5.7, 5.8, 6.1, 6.2 (all 64 bit)
Firmware Management Tool for the 9320 and 9720 Storage platforms.
Server and Storage hardware monitoring support for the 9320 Storage platform.
Failed node crash capture management Failed Node iLO Crash Management enables
customer to collect the core dump of the failed node (Segment Server) when Fusion Manager
initiates the failover. This functionality is not enabled by default.
Troubleshooting Guide enhancements
Manageability
Scheduling for antivirus scans
Languages 5
Rebalancing wizard
Segment evacuation wizard
Data tiering wizard
Replication wizard and health checks
NOTE: The 9th column (FFREE) which is your total available inode count per segment for the
original segments 66 Million per segment and that of the newer 64 bit segments of 1 billion per
segment. This segment mix and inode count does not negatively affect the operation of your file system
nor any applications.
Software snapshots
Upgrading pre-6.0 file systems for snapshots
To accommodate software snapshots, the inode format was changed in the 6.0 release. Consequently,
files used for snapshots must either be created on StoreAll OS software 6.0 or later, or the pre-6.0 file
system containing the files must be upgraded for snapshots. To upgrade a file system, use the
upgrade60 utility. For more information about the utility, see upgrade60 in the HP StoreAll Storage
CLI Reference Guide.
Fixes
Fixes in the 6.2.3 release
The firmware bundle includes new firmware version 1.40 for the X9720 storage controllers
(ExDS9100cc). Version 1.40 contains critical fixes and is a mandatory upgrade. After updating
the storage controllers to version 1.40, it may take 10-12 minutes for the storage controllers to
complete a reset and restart with the new firmware. During this time, issuing the command
hpsp_fmt --flashrec for the storage_ctlr component may report missing devices. Users
should wait until both controlls restart (with the new firmware) before performing any other
operations on these controllers.
Fixed an error counter issue (the number of errors did not match the number of corrections) in
ibdircheck. Now, the error counter is increased once (instead of twice) when ibdircheck
finds and fixes a problematic dentry.
ibdircheck no longer deletes a freshly-resolved migrated hard link. Previously, if a problematic
dentry was successfully resolved by lookup, ibdircheck did not check the dentry again and
the dentry may be accidentally deleted.
During inode creation, the inode is now checked to see if it is a replica. This prevents inode
creation from failing due to the inode quota limit being exceeded.
Fixes 7
Some SNMP managers, such as Netcool Omnibus Manager, received StoreAll SNMP traps with
an incorrect Specific Trap type and mismatched variable bindings.
The fix changes the SNMP Trap OID (OID of IBRIXGENERICEVENT) to .1.3.6.1.4.1.18997.2.0.1
and adds eventIndex to the variable bindings. Additionally, this fix includes a change to the
StoreAll MIB file, which requires the following configuration steps after installing the 6.2.3 patch:
1. On the CMS server, you must compile and manually register the StoreAll MIB file using HP
Systems Insight Manager 7.1:
a. Download ibrixMib.txt from /usr/local/ibrix/doc/ on the StoreAll server.
b. Rename the file to ibrixMib.mib.
c. In HP Systems Insight Manager, complete the following steps:
1) Unregister the existing MIB using the following command:
<BASE>\mibs>mxmib -d ibrixMib.mib
2) Copy the ibrixMib.mib file to the <BASE>\mibs directory and execute the
following commands:
<BASE>\mibs>mcompile ibrixMib.mib
<BASE>\mxmib -a ibrixMib.cfg
2. Once the new MIB file is successfully compiled and loaded, you must apply the
ibrixSnmpEvent.sql in HP Systems Insight Manager 7.1:
a. Download ibrixSnmpEvent.sql from usr/local/ibrix/doc/ on the StoreAll
server.
b. In HP Systems Insight Manager, perform the following steps:
1) Copy the ibrixSnmpEvent.sql file to <BASE>\config\preload\71\
addfiles\.
2) Execute the following command from CLI:
<BASE>\mxconfigrefresh
NOTE: For more information about loading the MIB, see the Compiling and customizing MIBs
chapter in the current Systems Insight Manager User Guide, which is available at: http://
www.hp.com/go/insightmanagement/sim/.
If HP Systems Insight Manager is shared between StoreAll clusters, you must install the 6.2.3 patch
and complete these steps on all StoreAll clusters.
8 Fixes
Running replication with antivirus scanning no longer corrupts ibrmode, which previously made
the scan appear not to occur.
When performing a segment evacuation on a shadow directory (due to a bad segment), the
evacuation no longer fails.
Previously when hard links were created on a StoreAll file system from the StoreAll Windows
Client, unmounting the file system might result in a failure. The fix ensures the failure no longer
occurs.
The file name length value was corrected in the Notify response packet to prevent the SMB client
from failing during create, change, or delete operations on files or folders.
The HP-SMB Srvsvc service no longer fails due to an excessive number of RPC connection
requests.
Access-based enumeration now works for all types of directory enumeration requests for SMB
clients.
After an upgrade on the StoreAll 9730, a simultaneous reboot of multiple servers could cause
the SAS expanders in MDS600 and D6000 disk enclosures to remain undiscovered. This resulted
in the hard disk drives not appearing and the disk enclosure failing to come online, even in a
degraded mode. This has been corrected with an upgrade to Virtual SAS Manager firmware
version 3.2.2.0.
After archiving files, segment evacuation no longer fails and an RO lock error no longer occurs
when accessing files on another node.
Significantly improved sequential read performance for HP-SMB (up to four times faster compared
to previous releases).
The HP-SMB service called lwiod would restart due to an assertion failure, which occurred when
a client attempted to recover a previously opened session after a disconnect/connect failure.
The HP-SMB share management utility called lwnet previously failed when run with the lwnet
share --set-info option. This occurred only when a domain user was used in the command,
and the node had been removed from the domain.
The HP-SMB service lwio failed when multiple clients accessed files or folders under a share and
simultaneously an application like PowerShell subscribed to event notifications on a shared folder.
When using the ibrix_avquarantine command, if the complete path length for the infected
file exceeded 200 characters, a crash occurred and the infected file was not moved.
Fixed the order of components during an upgrade so that creating archiving-enabled file systems
post-upgrade no longer requires a restart of the Fusion Manager component.
LUN discovery was corrected to handle multiple comma-delimited vendor storage names.
Fixed an issue that caused i_blocks and i_total_blocks to become out of sync when a directory
was being created.
Refresh stale dentry now displays when a misidentified dangler or a legitimate corrupted
directory condition occurs.
A missing attribute in the LDAP server user record now generates an error in the log, but no longer
results in a crash.
Wrapper scripts were modified to accept spaces in arguments.
The IBFS driver registry entry is now stored in memory to avoid the overhead of reading from the
registry.
Fusion Manager correctly redirects stdout and stderr messages to the log file, which ensures
that phases 2 and 3 of the ibrix_fsck command will complete.
The file name in the DCN response was changed to be the relative Windows file name (which is
the Windows server behavior) instead of the absolute UNIX file name.
Fixes 9
The default number of adm server threads was increased from 10 to 24. Previously, adm server
threads were used only for Fusion Manager and IAS management but many utilities now use adm
server threads.
A race condition could occur in the StoreAll kernel when a delegation thread was deallocated
without being properly synchronized with other threads. This occurred only in very rare
circumstances. The fix ensures that proper synchronization occurs before performing the thread
deallocation.
An electrical signal on the cache module is now synchronized before being analyzed. Previously,
electrical noise on this signal had been driving the supercapacitor charging logic into an undefined
state.
A deadlock in the switchs internal thread handling was eliminated.
The controller firmware now retries certain operations that were prematurely marked as failed.
A reset signal on the cache module was previously being cleared prematurely in brownout
scenarios; this has been corrected.
The firmware was corrected to ensure that drive heads are safely parked during a loss of power.
Previously, the switch firmware was miscounting SAS frame errors, resulting in a generic warning
being displayed on the SAS switch.
The Firmware Management Tool now displays a more informative error message when it is unable
to retrieve a list of file serving nodes from the Fusion Manager.
The firmware bundle includes disk drive firmware version HPD3 for disk drives MB1000FBZPL
and MB2000FBZPN. HPD3 prevents an infrequent condition that may occur when the WRITE
SAME command is used during RAID ARRAY parity initialization, which can result in unexpected
data on the hard drive.
Fixed an issue with StoreAll Linux clients running Red Hat Enterprise Linux 6.x in which the system
froze for five seconds during file editing.
10 Fixes
Request for a method for setting the pasv_address parameter in the vsftpd.conf file. This
parameter can now be set as follows:
CLI. Enter the following command, which will write the pasv_address parameter
(pasv_promiscuous=YES) into the sftpd configuration file:
ibrix_ftpshare -m ftp_share1 -c profile1 -I 10.9.60.61:21 -S
"allow_nat=true"
GUI. A new field named Allow NAT connections has been added to the FTP share modification
window.
Fusion Manager (FM) restarted a CFR job even if it is marked stopped by the user. Fusion
Manager has been modified so that it does not automatically restart a CFR job when the user
requests the CFR job to be stopped.
A soft lockup of the CPU was occurring during connection/delegation recovery.
The creation of directory tree quota names with a colon (:) in the name produced undesirable
results and it further impacted the creation of new "normal" directory tree quota names until the
entry with the colon character was removed. Users are now told in the HP StoreAll Storage File
System User Guide that they cannot create a directory tree quota containing a comma (,) or colon
(:) character.
When performing a "graceful" shutdown (ibrix_umount) of the StoreAll system for maintenance
reasons, the segment server hosting the root segment took 30 minutes per NFS export to timeout
and finally unmount.
When accessing StoreAll file systems through NFS, the client NFS access would hang intermittently.
When segments reached 7080% of capacity, clients were unable to write to an SMB share.
The quota current size would be reduced after running the rebalance task, even though files were
not modified or deleted. During movement of files across segments, there are intermediate replica
files that are created. Once data from an original file is synced with its replica, the original file
is deleted and the replica is promoted as the original file. These replicas have to be marked with
the directory quota ID, which is the same as that of the original master file. That was not occurring.
As the original file was deleted post replication, the directory quota accounted for it was getting
decremented. Since the newly created file was not marked with the directory quota ID, it was
never was accounted for. As a result, a reduction in quota usage was shown post replication.
The fix was to mark the replica created with the quota ID same as that of the original file, so any
data written into this file is accounted for against the set quota ID.
Data tiering could result in zero length files.
When a snapshot was taken, files could be deleted from the source directory.
Stack overflows were occurring, due to some of the CLI commands, such as ibrix_fs, requiring
a large buffer. The fix lessened the risk of a stack overflow.
Dangling dentries were created when mv and cp commands were killed during the file creation
operation.
The upgrade60.sh upgrade script was sometimes showing an invalid file system message in
environments where clusters were on a flat network.
A delay would occur between serial sequential file writes, about 2 to 20+ seconds between the
first file transfer and the next. This delay would be seen during the drag and drop of multiple 4
GB or larger files from Windows Explorer onto the StoreAll SMB share.
Checksum errors were being reported while accessing multiple database files from SUSE Linux
NFS Clients, which were reading multiple database pages (8 KB) into the NFS Client memory for
a SAP application. Although some of the data inside the 8 KB page was missing, when the same
page was re-read, the page data was completely intact and the checksum was correct. This
Fixes 11
situation only occurred when the read operations involved a cluster network hop to a remote FSN
on a remote couplet, and the issue has been resolved.
During a mkdir operation, the xdr_encode_netobj dereferencing NULL
xdr_netobj->data pointer message was displayed. The underlying issue has been resolved
so the message no longer appears.
During Directory Change Notification (DCN) processing, a serialization issue could cause the
thread to loop forever while traversing the list of consumers of the notification. The Linux kernel
would subsequently detect a soft lockup and log messages in regards to the situation would appear
in the kernel log (BUG: soft lockup CPU#<X> stuck for 10s!
[i_worker_X:<pid>]). The serialization issue has now been fixed.
There was a request to document file system mount information (specifically, the change to the
default mount options in 6.1 and the use of ibrix_mountpoint to set mount options). The man
pages for the ibrix_mount and ibrix_mountpoint commands and the HP StoreAll Storage
File System User Guide were updated.
The StoreAll Windows Client would crash when the client was built in debug mode.
The following fixes have been made for the 9730 Storage platform:
Storage could become unresponsive when multiple surface scans were launched on a single
physical disk. Interaction between background surface scans and performance has been
significantly improved.
The controller caching policy has been adjusted to significantly reduce the occurrence of
command timeouts and aborts when running heavy workloads with small block sizes.
Drive LEDs could go dark when an MDS600 storage enclosure I/O module was
removed. Drive LEDs are now refreshed in this scenario.
Sometimes MDS600 storage enclosure I/O modules were not visible to the storage controllers
after a FW flash of the I/O module. This has been corrected.
Improved resilience against SAS topology disruptions due to events such as booting of blades,
which could lead to missing paths to storage.
For a list of firmware fixes, see the FMT User Guide (contact HP Support for more information).
MDS600 and D6000 drive enclosure drawers are now reported by number: Drawer 1 for the
front-facing left drawer and Drawer 2 for the front-facing right drawer.
A memory leak was occurring in Fusion Manager.
For a list of fixes in the Storage Provisioning Tool (SPT), see the HP SPT User Guide (contact HP
Support for more information).
The var/log/message file was not getting rotated nightly.
Workarounds
This section contains workarounds for product situations that may occur:
12 Workarounds
http://saw.cce.hp.com/km/saw/view.do?docId=emr_na-c03466768
When creating a user NIC with a VLAN tag on a bond, running the ibrix_nic -c command
causes the parent bond to become degraded. The workaround is to determine the device that is
down and then bring up that device manually. For example, if bond0 is made up of eth4 and
eth5, use ifcfg to determine which device is down. Then use ifup to bring up the device,
such as ifup eth4.
Upgrades
The supported upgrade path from 5.6 or earlier is to first upgrade to 6.1. Before upgrading from
5.6 or earlier to 6.1, ensure that the NFS exports option subtree_check is the default export
option for every NFS export. The no_subtree_check option is not compatible with the StoreAll
OS software.
To add the subtree_check option, perform the following steps:
1. Unexport NFS exports:
ibrix_exportfs h <HOSTLIST> -p <CLIENT1:PATHNAME1>
2. Create NFS exports with the subtree_check option:
ibrix_exportfs f <FSNAME> -p <CLIENT1:PATHNAME1> -o subtree_check
NOTE: Multiple options can be specified using the o parameter by separating each option
by a comma (for example, -o rw,subtree_check).
If you plan to do an online upgrade, I/O must be occurring on a user virtual interface (VIF). Online
upgrades are only supported when I/O to shares is exclusively through User VIFs. When a VIF
is used, the server failover process sends the VIF to the running/active node. If I/O has occurred
through a non-virtual IP for the server, this I/O might need to be halted, and you might need to
reboot the server to proceed with an online upgrade.
After upgrading to 6.2.3, use the Firmware Management Tool (FMT) by entering the hpsp_fmt
fr server) command to determine if any firmware upgrades are required. If you are shown
the following message, proceed with the workaround:
ERROR: Could not recommend any flash action for one or more devices of
the following component type(s):
--------------------------------------------------------------------------------
Device :NIC
Code :0x1A0E
ERROR :Discovery failed for HP Embedded Network Adapter using
CP017785.scexe. Check log file
/var/log/hp/platform/fmt/firmware.log for more
information
--------------------------------------------------------------------------------
To resolve the issue:
Workarounds 13
1. Manually check your version by entering the following command:
ethtool -i eth0
The command displays the following:
driver: be2net
version: 4.1.402.6
firmware-version: 4.1.402.20
bus-info: 0000:02:00.0
Bootable ISO Download:
2. Verify the firmware-version is 4.1.402.20. If it is not, download a bootable ISO file
(OneConnect-Flash-4.1.402.20.iso) from the following link to perform the NIC
upgrade offline:
http://h20000.www2.hp.com/bizsupport/TechSupport/SoftwareDescription.jsp?lang=en&
cc=us&prodTypeId=3709945&prodSeriesId=4085948&prodNameId=4085929&
swEnvOID=54&swLang=8&mode=2&taskId=135&swItem=co-106538-1
The HTTP WebDAV share might not let you create a file or folder with the following message:
You dont have permission to do this action.
To resolve the issue, remount the WebDAV share by entering the following command:
net use * http://192.168.1.1/smita/
In this instance, the HTTP WebDAV share is 192.168.1.1/smita.
The automated offline upgrade appears confused by clusters with more than 10 nodes. This issue
results in the upgrade failing when it attempts to connect via SSH to a filename rather than to a
server.
To resolve the issue, move the Active Fusion Manager (FM) back to the initial node by using the
following process:
1. From the current Active FM, disable failover across the cluster by entering the following
command:
ibrix_fm -m nofmfailover -A
NOTE: The above command does not place the local/Active FM in nofmfailover mode.
val_scan_ended
The audit event types are in place to support this functionality in the future.
In the StoreAll REST API, the system::tier attribute is populated correctly when the file is first
created, or when it is first added to the metadata database if Express Query is enabled on an
existing file system. However, when the file system is moved to another tier by StoreAll file migration
utilities such as data tiering or rebalancer, that change does not yet update the system::tier
attribute available for query through the REST API.
Root segment evacuation fails on a file system that has express query and retention enabled.
Before running segment evacuation, disable Express Query by entering the ibrix_archiving
Workarounds 15
S command. You can re-enable Express Query after segment evacuation completes. To enable
Express Query, enter the ibrix_archiving s command.
In certain cases where files created by MDExport or ibrix_audit_reports are very large, in the
millions of files range, the MDImport tool cannot successfully import the files. To allow the files to
be imported, break up the output files into smaller files, on row (line) boundaries, and use the
MDImport tool to import each of the files.
Management console
If you need to make cluster configuration changes such as adding, extending, or deleting a file
system, be sure that at least two file serving nodes in the cluster are UP and are not in nofmfailover
mode. If a configuration change is necessary and only one node is UP, save a copy of /usr/
local/ibrix/tmp/fmbackup.zip after each configuration change is completed. In the case
where only one node is UP, this copy can be used later if that node goes down before any of the
other nodes come up. In general, before you bring up cluster nodes that have been down for
some time, ensure that the cluster already has an active Fusion Manager running.
When the active management console is moved to nofmfailover mode, a passive management
console will transition to active mode. Be sure that this transition is complete before you move the
previously active management console from nofmfailover mode to passive mode. (Use the
ibrix_fm -i command to check the mode of each management console.) If the passive
management console has not yet assumed active mode, the management console being moved
from nofmfailover to passive mode will become active again.
File system creation on the 9300 server fails from the Getting Started Wizard. No data loss or
corruption occurs, and you can still create the file system using the Create Filesystem Wizard or
its comparable CLI command (ibrix_fs -c -f...).
File system creation from the Management Console, not the CLI, might sometimes fail with the
message Exporting filesystem (NFS) failed. The file system was created correctly and
all the other steps in the process have completed.
After creating a file system, you can create an NFS share and export the file system using either
the GUI or CLI, as described in the HP StoreAll Storage File System User Guide.
Fusion Manager
Actively running tasks (Fusion Manager event-based) stop reporting status to the active Fusion
Manager (FM) if an FM restart occurred on the system doing the monitoring for that task.
To resolve the issue:
1. Manually delete the task by entering the following command:
ibrix_tasks -k -n <taskid> -F
2. If the FM process was stopped or restarted while a task was running, the common user
database (CUD) and the specific task coordinator (or other task specific process) were left
running. When FM is restarted, it is no longer aware of these processes. A new task of the
same type might not start because the rogue task is still running. Remove the rogue task by
manually stopping the coordinator/task specific processes and restarting the CUD.
Under some circumstances the flashing of some hardware components can fail. See the HP FMT
User Guide for details (contact HP Support for more information about this guide).
If Fusion Manager (FM) is taking a long time to start and the logging for FM is showing APPARENT
DEADLOCK messages, purge the event database.
To purge the event database:
16 Workarounds
1. Enter the following CLI command:
ibrix_event -p [-o DAYS_COUNT]
If you want to change the number of days the event database is purged, include the -o
DAYS_COUNT option when you enter the command.
Events are removed starting with the oldest events. The default is to remove the oldest seven
days of messages. Purge the event database periodically to prevent it from becoming too
large.
2. After the events are purged, create a new FM Store archive by entering the following
command:
ibrix_fm -B
3. Wait for a few seconds to let the Store archive propagate through the cluster and then restart
the active FM.
The credentials management service (hpsp_credmgmt) will not allow adding a credential for a
new MSA storage array after the credential database is created. The workaround is to install the
credentials manually.
To install the credentials manually:
1. Enter the hpsp_couplet_info -i command to get a list of UUID for attached MSA
enclosures. Each enclosure is on a separate line, as shown in the following example of the
command and its output:
[root@ib25-02s1 security]# hpsp_couplet_info -i
500c0ff1390c0000
2. Install the credentials for a single MSA by running the following four commands once for
each MSA that the hpsp_couplet_info command lists in step 1. Replace
<MSA_SAS_BASE_ADDRESS> with one of the UUIDs as provided by the output of the
hpsp_couplet_info command.
hpsp_credmgmt --update-cred -s
couplet:couplet/array/<MSA_SAS_BASE_ADDRESS>/mgmtport_a/rw
--cred-type upwpair --cred-username manage --cred-password '!manage'
hpsp_credmgmt --update-cred -s
couplet:couplet/array/<MSA_SAS_BASE_ADDRESS>/mgmtport_a/ro
--cred-type upwpair --cred-username monitor --cred-password
'!monitor'
hpsp_credmgmt --update-cred -s
couplet:couplet/array/<MSA_SAS_BASE_ADDRESS>/mgmtport_b/rw
--cred-type upwpair --cred-username manage --cred-password '!manage'
hpsp_credmgmt --update-cred -s
couplet:couplet/array/<MSA_SAS_BASE_ADDRESS>/mgmtport_b/ro
--cred-type upwpair --cred-username monitor --cred-password
'!monitor'
Workarounds 17
3. To verify your installation of the credentials, list the credentials by entering the following
command:
hpsp_credmgmt --get-cred -s couplet:couplet/* --output-sl
The command displays the credentials for your system. Your output would differ from the
following example to reflect your system configuration:
10.125.2.221|monitor|!monitor|couplet:couplet/array/500c0ff1390c0000/mgmtport_b/ro
10.125.2.221|manage|!manage|couplet:couplet/array/500c0ff1390c0000/mgmtport_b/rw
10.125.2.220|monitor|!monitor|couplet:couplet/array/500c0ff1390c0000/mgmtport_a/ro
10.125.2.220|manage|!manage|couplet:couplet/array/500c0ff1390c0000/mgmtport_a/rw
Hardware monitoring will continue to show a storage cluster in a DEGRADED state when an MSA
snapshot LUN is removed due to the reserve LUN getting full or the user removing the block
snapshot from a StoreAll 9320 Storage platform. To resolve this issue, stop the monitoring process
via the /etc/init.d/hpspmon stop command. The monitoring process will be restarted
automatically and will reset the statuses.
Issues with just the StoreAll 9320 Storage Platform:
If an MSA I/O module is MISSING the status of its sensors (Capacitor, Temp) will appear
as UNKNOWN. After the I/O module is replaced there is a period of one minute during
which status for all sensors in the array is unavailable. The statuses will be corrected in the
next monitoring polling cycle.
If there is a hardware event involving a fan in a D2700 drive enclosure then the location of
the fan will incorrectly state that it is part of the enclosures power supply. The fan is a separate
FRU in the D2700 drive enclosure.
Disconnected SAS cables between D2700 small form factor drive enclosures are not detected
by the hardware. Monitoring software will not report lack of redundancy in connectivity to
drive enclosures that are connected behind the first drive enclosure. If the customer restarts
the other storage controller that has access to the enclosure or removes the other cable to the
enclosure loss of access to storage will occur.
Spurious failure indications and potential system crashes can occur if storage controller or
hard drive firmware is flashed while monitoring is active. Run the hpspmonControl
--pause command on both servers in the couplet before performing the flash operation.
Then, run the hpspmonControl --resume command on both servers in the couplet after
the flash operation is complete.
Disconnecting the cable from the HBA to an MSA or rebooting one or both of the MSA
storage controllers while hardware monitoring is active can result in unexpected hardware
events.
Port monitoring of the HP SC08Ge HBA (3GB) is not functional when the HP SNMP agents
are not running.
On an IBRIX X9720 or StoreAll 9730 Storage Platform, when a blade is reseated in the C7000
chassis, the On-Board Administrator (OA) might allow the blade to be manually powered on
18 Workarounds
before Virtual Connect (VC) profile is applied. The server might come up with no networking and
without any error in OA. The VC sometimes complains.
The Active Fusion Manager (FM) might lose track of actively running tasks if the FM on the server
where the specific task coordinator is running is restarted. This situation can happen if the server
has failed over, restarted or if the Fusion Manager itself on that server has restarted.
The following tasks are impacted by this issue:
OnlineQuotacheck
Migrator
SnapReclamation
DataValidation
CaseInsensitive
The following symptoms might be observed:
The task may appear to be hung in the Fusion Manager task display output because it will
only display the last status it received.
Another task of the same type may fail to start on the same file system due to leftover rogue
processes from the previous task.
To resolve this issue:
1. Enter the following command:
ibrix_task k n <taskid> -F
2. Manually kill the specific coordinator process related to rogue task.
See the following table for information on how to kill each task.
Hung Task Enter the following command on one line to kill the specific coordinator process
NOTE: If there is a worker on each server that owns a segment of the file system, enter the
following command for each worker.
/usr/local/ibrix/bin/migrator_daemon --serviceid=4100 --port=9174 -f
ifs2 --logflag=6 -L /usr/local/ibrix//log/ -j Migrator_58-ifs2-worker1
CaseInsensitive /usr/local/ibrix/bin/dir_tree_case_insensitive -p -j 51
Workarounds 19
SMB
SMB and StoreAll Windows clients cannot be used together because of incompatible AD user to
UID mapping. You can use either SMB or StoreAll Windows clients, but not both at the same time
on the cluster.
The HPSMB server does not support connections from Linux SMB clients. The workaround is to
use NFS for Linux.
SMB DCN is supported on Windows File Explorer with the following exception:
The recursive notification flag in the notify request is ignored. Notifications are delivered only
for one directory level.
If SMB is joined to a domain before introducing LDAP ID mapping or changing the domain or
the provision mode, you should clear the name cache on each node. This step prevents unexpected
results such as the corruption of a name cache entry when the new configuration is in place. To
clear the cache, run the following command:
/opt/likewise/bin/lw-lsa ad-cache --delete-all
When the Microsoft Windows Share Management interface is used to add a SMB share, the
share path must include the StoreAll file system name. The Browse button on the MMC cannot be
used to locate the file system. Instead, enter the entire path, such as C:\data\.
The StoreAll management console GUI and CLI allow only 9000 file systems and directories to
be exported as SMB shares. However, the Microsoft Windows Share Management interface
allows you to create an SMB share that is not on a StoreAll file system. Although the share will
be available from the file serving node to which Windows Share Management was connected,
it will not be propagated to the other file serving nodes in the cluster.
The ibrix_localusers -i <user information> command fails if the user information
includes commas. To enter commas in the user information, use the management console GUI
instead of the CLI.
When you use the Windows security tab to add local users or groups to a security ACL on an
SMB file (for either file or share-level permissions), you typically specify the user to add as either
a DOMAIN\username or a MACHINE\username. On StoreAll systems, local users are displayed
as LOCAL\username, and it may seem like you should specify LOCAL\username in the Add
dialog box in Windows. However, in this case, the Windows client cannot interpret LOCAL.
Instead, specify the machine name of the server. For example, to add LOCAL\user1 to an ACL
on an SMB file shared out by serverX, specify serverX\user1 in the Add dialog box on the
security tab. If you later use the Windows security tab to look at this ACL, the server name will
have been replaced by LOCAL (the SMB server performs this remapping to ensure that local users
are symmetric between all servers in the cluster, and are not specific to any one machine name
in the cluster.)
When joining an SMB domain, the $ character cannot be used in passwords unless it is escaped
with a slash (\) and enclosed in single quotes (' '). For example:
ibrix_auth -n IB.LAB -A john -P 'password1\$like'
If you are using a Windows Vista client and running more than a single copy of Robocopy from
that client, a hang is possible. The work around is to disable the SMB2 protocol on all file serving
nodes. Complete these steps:
1. Disable the SMB2 protocol on all nodes:
ibrix_cifsconfig -t -S SupportSmb2=0
2. Restart the SMB service on all nodes:
ibrix_server -s -t cifs -c restart
20 Workarounds
You may also need to restart the Windows client, as the original negotiated protocol, SMB2, may
be cached by the client. Restarting the client renegotiates the protocol back to SMB1.
Be sure to remove Active Directory users from the StoreAll share admin list before removing them
from Active Directory. If you remove an Active Directory user from Active Directory before removing
the user from the StoreAll share admins list, an error is reported when you attempt to change the
share admins list. If you are seeing errors from this situation, rejoin Active Directory and remove
all share admins. For example:
ibrix_auth -n ib.lab -A administrator@ib.lab -P fusion -S "share
admins="
Then run ibrix_auth again to specify the new list of share admins:
ibrix_auth -t -S "share admins=[ib\Administrator]"
Do not use spaces in share names. The GUI cannot create a share if there is a space in the name.
SMB clients have a 45 second timeout on every SMB request they make. If a request's timer
expires before a response comes, the client disconnects and terminates the SMB connection with
this error message: The specified network name is no longer available. After
this, the client must reconnect to the SMB server. If you are seeing this message from clients
connected to a busy server, they may be encountering timeouts from overload. For more information
about SMB client timeouts, see http://support.microsoft.com/kb/102067.
LDAP
When configuring LDAP on the GUI, double quotes are not supported in any fields of the LDAP
configuration window, including the password field. Also, escaped slashes are not supported. If
you enter \\; it will be parsed literally as \\ instead of \. Also, the ibrix_ldapconfig
command supports double quotes only at the beginning and end of a value. For example:
ibrix_ldapconfig -a -h 10.30.239.174 -b "cn=Manager,dc=ldapx,dc=net"
-p "mypassword "
The command does not support escaped slashes.
LDAP ID mapping
When LDAP ID mapping is configured with the GUI or CLI, the following fields are not validated
to ensure successful communication with the LDAP server: LdapBase, LdapServer, LdapUserDn,
LdapLapMaxWaitTime, LdapMaxEntries fields. It may be helpful to perform a manual ldapsearch
test with the intended configuration values to assure success when configuring LDAP ID mapping.
The LDAP ID mapping features depends on successful communication with the configured LDAP
server.
The openldap-clients-<version> RPM is required to perform ldapsearch commands
from the command line. This RPM is available on the OpenLDAP site.
Increasing the values of the LdapMaxWaitTime and LdapMaxEntries can affect the performance
of name resolutions for SMB and non-SMB users in large LDAP directory environments. The AD
name cache protects against excessive LDAP searches. The default value for these parameters is
10. The range of values is 0 (unlimited) or 132767.
Block snapshots
Snapshot creation may fail while mounting the snapshot. The snapshot will be created successfully,
but it will not be mounted. Use the following command to mount the snapshot manually:
ibrix_mount -f <snapshotname> -m /<snapshotname>
Quotas are disabled on block level snapshots (for example, MSA2000 snapshots) and the quota
information from the origin file system is not carried to the block level snap file system. Block level
Workarounds 21
snapshots are temporary file systems that are not writable. Users should not query quota information
against block level snap file systems.
After the initial creation of a snapshot, it can take 4 to 6 minutes to mount the snapshot.
Block snapshots are not created when a volume group is mapped to more than one logical volume.
Remote Replication
Remote replication might fail to transfer Windows Access Control Lists (ACLs) if cross-protocol
ACL synchronization is enabled on the source cluster nodes, but not on the target cluster nodes.
Before starting the replication, ensure that cross-protocol ACL synchronization is enabled on all
source and target cluster nodes.
When remote replication is running, if the target file system is unexported, the replication of data
will stop. To ensure that replication takes place, do not unexport a file system that is the target
for a replication (for example, with ibrix_crr_export -U).
Remote replication will fail if the target file system is unmounted. To ensure that replication takes
place, do not unmount the target file system.
When continuous remote replication is used and file serving nodes are configured for High
Availability, you will need to take the following steps following failure of a node:
1. Stop continuous remote replication.
2. After the migration to the surviving node is complete, restart continuous remote replication
to heal the replica.
If these steps are not taken, any changes that had not yet been replicated from the failed node
will be lost.
No alert is generated if the continuous remote replication target becomes unavailable. Confirm
the connection to the target system by issuing a ping command and by inspecting
ibrcfrworker.log.
Sparse files on the source file system are replicated unsparse on the target. That is, all blocks
corresponding to the file size are allocated on the target cluster. Consequently, if the target file
system is the same size as the source file system, remote replication can fail because there is no
space left on the target file system. To work around this situation, if the source system contains
large sparse files, be sure that the target file system is larger than the source file system, and large
enough to fit all files in an unsparsed manner.
The mountpoint /mnt/ibrix is reserved for remote replication. Hiding or blocking this mountpoint
by mounting anything over the parent /mnt will prevent Run Once replication from working at
all, and the initial domain scan of Continuous replication will fail.
When a replication task encounters a file on the target system that was added or changed outside
of a previous replication, the replication task stops and does not replicate the remaining files from
the source to the target.
If the target file system is unmounted while an inter-cluster or intra-cluster CRR task is running, the
CRR task will not work properly. Before unmounting a file system, it is a best practice to verify
that the file system is not the target for a CRR task. If the file system is the target for a CRR task,
stop the task before unmounting the file system.
22 Workarounds
Data retention and validation
The ibrix_reten_adm command fails if the date string specified with -e contains spaces. As
a workaround, use the following command to enter date strings containing spaces:
/usr/local/ibrix/sbin/ibr_reten_adm -e expire_time -f FSNAME -P PATHLIST
The ibrix_vs_snap command cannot delete a block snapshot file system that is enabled for
data retention. Instead, use the ibrix_fs command with the -R option. For example:
ibrix_fs -d -f block_snap_ifs2_1 -R
When the GUI or ibrix_reten_adm -r is used to remove a legal hold from a file, it displays
invalid errors about autocommit; however, the command completes successfully. You can ignore
the error messages.
Directories are not supported in the path list for the ibrix_reten_adm command. However,
the command does not display the correct error message when the path list includes a directory.
PDF reports do not display on Internet Explorer with Adobe Reader 10.1.2 installed. Reports
display properly with earlier versions of Adobe Reader or with other supported browsers.
The ibrix_reten_adm -e command fails for WORM files that are not retained. If a WORM
file is not retained (either it has never been retained, or it was retained but its retention period
has expired), use the touch command to set its last accessed time (atime) forward to the desired
expiration time. The file will then be retained until that time.
PDF reports cannot be generated from the GUI when using Windows Internet Explorer 8. To
display the PDF, use one of these options:
On the browser, right-click in the white border surrounding the grey area, and select View
Source. Locate the src= line in the source. For example:
src="/fusion/reporting/retention/retention_ifs1.pdf#scrollbar=0"
width='100%' height='100%'/>
Prepend https://<hostname> to the path specified in the source and enter that URL in
the browser. Using the previous example, you would enter:
https://<hostname>/fusion/reporting/retention/retention_ifs1.pdf
Log into the node specified as the hostname in the URL and go to the directory /usr/local/
ibrix/reports/output/retention, which contains the report.
Antivirus support
If McAfee is used as a virus scan engine, when virus definitions are updated on the virus scan
engine, the updated definitions are not sent automatically to the StoreAll servers. To correct this
condition, request hotfix HF692368 from McAfee.
For performance reasons, when an NDMP backup is in progress, Antivirus scans are not triggered
for the files being backed up. The StoreAll server excludes the process corresponding to NDMP
from scanning any files on a cluster with Antivirus enabled.
HP recommends that you disable Antivirus on the file system when a non-NDMP backup is in
progress. StoreAll cannot exclude a non-NDMP backup process from scanning files during a
backup. Disabling Antivirus prevents scans for files being backed up, which improves backup
performance.
When WebDAV is enabled, the internal WebDAV files in the root of the file system are scanned.
To exclude all .lock/.pag files in the .DAV and .webdav directories from scanning, complete
these steps:
1. After the WebDAV share is created, unset the immutable bits on .webdav directory:
# chattr -i .webdav/
Workarounds 23
2. Run the following command to exclude all files in the .webdav directory:
# ibrix_avconfig -a -E -f FS1 -P /FS1/.webdav/ -x '*'
3. Reset the immutable bit on the .webdav directory:
# chattr +i .webdav/
4. Exclude all files in the .DAV directory for each HTTP/WebDAV share created:
# ibrix_avconfig -a -E -f FS1 -P /FS1/httpshare/.DAV/ -x '*'
Segment evacuation
The segment evacuator cannot evacuate segments in a READONLY, BROKEN or UNAVAILABLE
state.
/usr/local/ibrix/log/ibrixcollect/ibrixcollect.py.log
Email notifications do not include information about failed attempts to collect the cluster
configuration.
In some situations, ibrix_collect successfully collects information after a system crash but
fails to report a completed collection. The information is available in the /local/ibrixcollect/
archive directory on one of the file serving nodes.
The ibrix_collect command supports a maximum collection size of 4 GB. If the size of the
final collection exceeds 4 GB, the collection fails. You must either:
Delete the excess older logs from each node to reduce the size.
Or
Manually collect the individual collection, which is stored on each node in the following
format:
/local/ibrixcollect/<node_name>_<collection_name>_<time>.tgz
24 Workarounds
components to toggle from Up, to Stale, and back to Up. Common causes of failures in the
monitoring system include:
Reboot of a file serving node
Network connectivity issues between the management console and a file serving node
Resource exhaustion on a file serving node (CPU, RAM, I/O or network bandwidth)
While network connectivity and resource exhaustion issues should be investigated, they can occur
normally due to heavy workloads. In these cases, you can reduce the frequency at which
vendorstorage components are monitored by using the following command:
ibrix_fm_tune -S -o vendorStorageHardwareStaleInterval=1800
The default value of this command is 900; the value is in seconds. A higher value reduces the
probability of all components toggling from Up to Stale and back to Up because of the conditions
listed above, but will increase the time before an actual component failure is reported.
SNMP
When the StoreAll SNMP agent is configured, do not restart snmpd. If the snmpd restart
command is run on the active node, you will need to take one of these steps:
If Phone Home was configured, reconfigure it.
If Phone Home was not configured, reconfigure the StoreAll SNMP agent.
Workarounds 25
service snmpd restart
On 9320/X9720 systems, the virtual SAS manager and VC connect manager must be configured
manually for HP Support.
On HP SIM 7.1, 9320/9300 nodes are discovered as HP 9320 NetStor FSN if MSA storage
is registered on the cluster nodes. This does not affect sending alerts to HP Support.
ibrix_fsck command
The log files for user-initiated tasks (such as running ibrix_fsck and migrating or rebalancing
segments) are not rotated automatically. To control the space used by these log files, you will
need to maintain them manually. The log files are located in /usr/local/ibrix/log.
After running a tiering, rebalance or segment evacuation job, a phase 1 of the ibrix_fsck
command might display a message regarding bad replica inodes. The following is an example
of the message from a phase 1 ibrix_fsck log:
Entry 'F000000060000FEAF.DAT' in
/.ibrixShadowRoot/s_1/g_367/i_100b335b7_edd68d01/D00000006000005FE
(11812038) has bad replica inode 17877133.
This message indicates that a replica inode was deleted from the file system by a user application,
but the directory entry (DENTRY) pointing to the replica has not been removed. There is not an
issue with the file system, and this situation does not impact file system operations. The DENTRY
that points to the deleted replica can be removed by running a phase 1 ibrix_fsck in corrective
mode. See the HP StoreAll Storage CLI Reference Guide for more information.
Windows Client
If the StoreAll Windows Client is registered with Fusion Manager already, clicking on Start Service
After registration under the Registration tab displays a Response Code =400 error. Click the
Recover Registration option instead.
Reading of snaps as memory mapped files is not supported. Undesired results can occur if a snap
file is used as memory mapped file while the current version of the file is being modified by a
different instance.
General
The ibrix_pv -a -o mpath command does not recognize a multipath device. The command
works properly when an argument is added to ignore the standard SCSI devices. Execute the
command as follows, specifying the path to your device.
ibrix_pv -a -o accept:/dev/mapper/mpath10,~standard
The ibrix_lv -d -s <range of Segments> command sometimes displays a failed
command message. Although the segments for which the command fails are still listed by the
ibrix_lv -l command, they are successfully removed from the underlying Operating System
LVM view. This issue is a timing condition for large file system delete operations where the delete
command times out while waiting for a response from the storage. Although the deletion was
successful the command reports it as failed and StoreAll preserves the associated volume group
(ivg) and logical volume (ilv). This issue has been observed on the StoreAll 9730.
To resolve this issue, rerun the logical volume delete command using the -X override option, for
example: ibrix_lv -d -s <ilv#> X
The override option forces the removal of the StoreAll Logical Volume record from the StoreAll
database.
26 Workarounds
When using the ibrix_vs -a command during file system extension, the command fails on a
combined cluster configuration (such as 9730 and 9320) with the following error message:
Lun already set for physical volume dxx can two vendorstorages see the
same luns?
Contact HP Support for assistance.
You might not be able to mount a file system after an unmount if you:
1. Stop the metabox.
2. Evacuate the segment.
3. Failover of a server (ibrix_server -f -h <server_name>)
4. Unmount the file system.
To resolve the issue, restart the affected server before unmounting the file system.
If the ibrix_auth command fails after a QR restore of a server in a cluster with a message
similar to:
ibrix_auth -n IBRQA1.HP.COM -A administrator@ibrqa1.hp.com -P password
-h hostnameX
Iad error on host hostnameX failed command (<HIDDEN_COMMAND>) status
(1) output: (Joining to AD Domain:IBRQA1.HP.COM With Computer DNS Name:
hostsnameX.ibrqa1.hp.com )
Verify the contents of the /etc/resolve.conf file is not empty. If it is empty, copy the contents
of the /etc/resolve.conf file on another server to the empty resove.conf file.
When configuring automated failover with the HA wizard for the X9720, you must create the
ibrixpwr iLO user on each node before the cluster HA can be fully functional. Enter the
following command on each cluster node to create an iLO user with the user name ibrixpwr
and the password hpinvent:
ibrix_ilo c u ibrixpwr p hpinvent
HP-OpenIPMI does not work with the HP-Health service. For example, Proliant health check tools
such as SMH and SIM and hpasmcli commands such as SHOW POWERSUPPLY do not report
errors. This situation occurs because StoreAll OS software requires standard RHEL IPMI. Remove
the HP-OpenIPMI RPM (execute rpm -e HP-OpenIPMI), and then start the standard RHEL IPMI
(execute /etc/init.d/ipmi start). The standard RHEL IMPI will now start automatically
when the server is booted.
During server migration or failover, certain cluster events will be reported as alerts. These events
are expected and normal, and are reported temporarily as a server is failing over to another
server.
Node failover does not occur when a node has a complete loss of power (for example, removing
power cords or pulling a blade from a chassis). Do not test high availability in this manner.
NFS locks may return an error code (37) after an unlock operation even though the lock was
correctly released. If other results are normal, you can ignore this error code.
On 9730 systems, the iLO can fail to power up a server when directed to do so from the iLO web
page or the iLO's Integrated Remote Console. To correct this condition, reset the iLO processor.
This can be done from the iLO web UI; select Reset on the Diagnostics page.
After a server is rebooted following a QR install and before the server is configured, an error
message reports that the SRVSVCD service failed to load. The failure occurs because the service
requires that the network subsystem be configured. The error message can be safely ignored.
After the network is configured, restart the SMB service.
Workarounds 27
Installation instructions
New installations
HP StoreAll OS software is preinstalled on supported devices. If you need to reinstall the software, see
the administrator guide for your storage system.
Upgrades
The upgrade procedure is provided in the administrator guide for your storage system. Contact HP
Support for assistance with the procedure.
Compatibility/Interoperability
Note the following:
Every member of the cluster must be running the same version of StoreAll OS software.
The cluster must include an even number of file serving nodes.
Documentation feedback
HP is committed to providing documentation that meets your needs. To help us improve the
documentation, send any errors, suggestions, or comments to Documentation Feedback
(docsfeedback@hp.com). Include the document title and part number, version number, or the URL when
submitting your feedback.
28 Installation instructions