HU01866 |
SVC |
HIPER
|
A faulty PSU sensor, in a node, can fill the sel log causing the service processor (BMC) to disable logging. If a snap is subsequently taken, from the node, a timeout will occur and it will be taken offline. It is possible for this to affect both nodes in an I/O group
(show details)
Symptom |
Loss of Access to Data |
Environment |
SVC systems using SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
7.7.1.9 |
System Monitoring |
HU01767 |
All |
Critical
|
Reads of 4K/8K from an array can under exceptional circumstances return invalid data. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8.0 or earlier |
Trigger |
None |
Workaround |
None |
|
7.7.1.9 |
RAID, Thin Provisioning |
HU01771 |
SVC, V7000 |
High Importance
|
An issue with the CMOS battery in a node can cause an unexpectedly large log file to be generated by the BMC. At log collection the node may be taken offline
(show details)
Symptom |
Loss of Redundancy |
Environment |
SVC & V7000 systems running v7.8 or later |
Trigger |
Node CMOS battery issue |
Workaround |
None |
|
7.7.1.9 |
System Monitoring |
HU01445 |
SVC, V7000 |
Suggested
|
Systems with heavily used RAID-1 or RAID-10 arrays may experience a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
SVC & V7000 systems running v7.7 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.9 |
|
HU01624 |
All |
Suggested
|
GUI response can become very slow in systems with a large number of compressed and uncompressed volume
(show details)
Symptom |
None |
Environment |
Systems that have a large number of volumes (1000+) including some that are compressed |
Trigger |
None |
Workaround |
None |
|
7.7.1.9 |
Graphical User Interface |
HU01628 |
All |
Suggested
|
In the GUI on the Volumes page whilst using the filter function some volumes entries may not be displayed until the page has completed loading
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
Enter a filter string on the Volumes page |
Workaround |
None |
|
7.7.1.9 |
Graphical User Interface |
HU01664 |
All |
Suggested
|
A timing window issue during an upgrade can cause the node restarting to warmstart stalling the upgrade
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.9 |
System Update |
HU01687 |
All |
Suggested
|
For volumes by host, ports by host and volumes by pool pages in the GUI when the number of items is greater than 50 then the item name will not be displayed
(show details)
Symptom |
None |
Environment |
Systems running v7.7 or v7.8 |
Trigger |
More than 50 hosts/pools |
Workaround |
Use CLI |
|
7.7.1.9 |
Graphical User Interface |
HU01698 |
All |
Suggested
|
A node warmstart may occur when deleting a compressed volume if a host has written to the volume minutes before the volume is deleted
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with compressed volumes |
Trigger |
Write I/O received by compression code for a volume that has been deleted |
Workaround |
Remove volume host mappings and leave an interval of time (90mins) before removing the volume |
|
7.7.1.9 |
Compression |
HU01730 |
SVC |
Suggested
|
When running the DMP for a 1046 error the picture may not indicate the correct position of the failed adapter
(show details)
Symptom |
None |
Environment |
DH8 systems |
Trigger |
Run DMP for 1046 error |
Workaround |
Be aware of which adapter is failed and do not use the picture in the GUI |
|
7.7.1.9 |
GUI Fix Procedure |
HU01763 |
SVC |
Suggested
|
A single node warmstart may occur on a DH8 config node when inventory email is created. The issue only occurs if this coincides with a very high rate of CLI commands and high I/O workload on the config node
(show details)
Symptom |
Single Node Warmstart |
Environment |
DH8 systems running v7.7.1 or later |
Trigger |
High CLI workload and I/O workload at same time as inventory email |
Workaround |
None |
|
7.7.1.9 |
Command Line Interface, System Monitoring |
HU01706 |
All |
HIPER
|
Areas of volumes written with all-zero data may contain non-zero data. For more details refer to this Flash
(show details)
Symptom |
Incorrect data read from volume |
Environment |
Systems running 7.7.1.7 or 7.8.1.3 |
Trigger |
See Flash |
Workaround |
None |
|
7.7.1.8 |
|
HU00744 (reverted) |
All |
Suggested
|
This APAR has been reverted in light of issues with the fix. This APAR will be re-applied in a future PTF
|
7.7.1.8 |
Cache |
HU01239 & HU01255 & HU01586 |
All |
HIPER
|
The presence of a faulty SAN component can delay lease messages between nodes leading to a cluster-wide lease expiry and consequential loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with 16Gb HBAs |
Trigger |
Faulty SAN hardware (adapter, SFP, switch) |
Workaround |
None |
|
7.7.1.7 |
Reliability Availability Serviceability |
HU01505 |
All |
HIPER
|
A non-redundant drive experiencing many errors can be taken offline obstructing rebuild activity
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Drive with many errors |
Workaround |
Bring drive online |
|
7.7.1.7 |
Backend Storage, RAID |
HU01646 |
All |
HIPER
|
A new failure mechanism in the 16Gb HBA driver can under certain circumstances lead to a lease expiry of the entire cluster
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with 16Gb HBAs |
Trigger |
Faulty SAN hardware (adapter; SFP; switch) |
Workaround |
None |
|
7.7.1.7 |
Reliability Availability Serviceability |
HU01267 |
All |
Critical
|
An unusual interaction between Remote Copy and FlashCopy can lead to both nodes in an I/O group warmstarting
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using GMCV |
Trigger |
There may be a number of different scenarios that trigger this issue. Here is one: With GMCV running; Stop GMCV with -access (i.e. svctask stoprcconsistgrp -access); Secondary site becomes primary site and vice versa; Start a FlashCopy on the now primary site; Stop FlashCopy restore before it completes (with -force); Current primary becomes exposed to the issue; Reinstate GMCV (primary becomes secondary again); Stop GMCV with -access again; Both nodes warmstart. |
Workaround |
None |
|
7.7.1.7 |
Global Mirror with Change Volumes |
HU01416 |
All |
Critical
|
ISL configuration activity may cause a cluster-wide lease expiry
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Reliability Availability Serviceability |
HU01490 |
All |
Critical
|
When attempting to add/remove multiple IQNs to/from a host the tables that record host-wwpn mappings can become inconsistent resulting in repeated node warmstarts across I/O groups
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with iSCSI connected hosts |
Trigger |
- A addhostport command with iqn2 and iqn1 (where iqn1 is already recorded) is entered;
- This command attempts to add iqn2 but determines that iqn1 is a duplicate and the CLI command fails;
- Later whenever a login request from iqn2 is received internal checking detects an inconsistency and warmstarts the node
|
Workaround |
Do not use multiple IQNs in iSCSI add/remove commands |
|
7.7.1.7 |
iSCSI |
HU01519 |
V7000 |
Critical
|
One PSU may silently fail leading to the possibility of a dual node reboot
(show details)
Symptom |
Loss of Access to Data |
Environment |
Storwize V7000 Gen 1 systems |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Reliability Availability Serviceability |
HU01528 |
SVC |
Critical
|
Both nodes may warmstart due to Sendmail throttling
(show details)
Symptom |
Loss of Access to Data |
Environment |
V9000 systems |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
|
HU01549 |
All |
Critical
|
During a system upgrade HyperV-clustered hosts may experience a loss of access to any iSCSI connected volumes
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.5 or earlier with iSCSI connected HyperV clustered hosts |
Trigger |
Upgrade to v7.6 or later |
Workaround |
None |
|
7.7.1.7 |
System Update, iSCSI |
HU01572 |
All |
Critical
|
SCSI 3 commands from unconfigured WWPNs may result in multiple warmstarts leading to a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with iSCSI connected hosts |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
iSCSI |
HU01635 |
All |
Critical
|
A slow memory leak in the host layer can lead to an out-of-memory condition resulting in offline volumes or performance degradation
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Hosts, Performance |
IT20627 |
All |
Critical
|
When Read-Intensive drives are used as quorum disks a drive outage can occur. Under some circumstances this can lead to a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Using Read-Intensive drives as quorum disks |
Workaround |
None |
|
7.7.1.7 |
Quorum |
HU00762 |
All |
High Importance
|
Due to an issue in the cache component, nodes within an I/O group are not able to form a caching-pair and are serving I/O through a single node
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.3 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Cache |
HU01428 |
V7000, V5000, V3700, V3500 |
High Importance
|
Scheduling issue adversely affects performance resulting in node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Storwize systems running v7.7.1 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Reliability Availability Serviceability |
HU01477 |
V7000, V5000, V3700, V3500 |
High Importance
|
Due to the way enclosure data is read it is possible for a firmware mismatch between nodes to occur during an upgrade
(show details)
Symptom |
Loss of Redundancy |
Environment |
Storwize systems running v7.8 or later |
Trigger |
Upgrade from v7.8.0 to v7.8.1 |
Workaround |
Check for dependent volumes. If there are none then manually load enclosure firmware from service mode |
|
7.7.1.7 |
System Update |
HU01488 |
V7000, V5000 |
High Importance
|
SAS transport errors on an enclosure slot have the potential to affect an adjacent slot leading to double drive failures
(show details)
Symptom |
Loss of Redundancy |
Environment |
Storwize V5000 and V7000 systems running v7.7 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Drives |
HU01506 |
All |
High Importance
|
Creating a volume copy with the -autodelete option can cause a timer scheduling issue leading to node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All systems |
Trigger |
None |
Workaround |
Do not use the -autodelete option when creating a volume copy |
|
7.7.1.7 |
Volume Mirroring |
HU01569 |
SVC |
High Importance
|
When compression utilisation is high the config node may exhibit longer I/O response times than non-config nodes
(show details)
Symptom |
Performance |
Environment |
SVC systems using compression |
Trigger |
High compression workloads |
Workaround |
None |
|
7.7.1.7 |
Compression |
HU01579 |
All |
High Importance
|
In systems where all drives are of type HUSMM80xx0ASS20 it will not be possible to assign a quorum drive
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with drives of type HUSMM80xx0ASS20 |
Trigger |
Attempt to assign drive type as quorum |
Workaround |
Manually assign a different drive type as quorum |
|
7.7.1.7 |
Drives, Quorum |
HU01609 & IT15343 |
All |
High Importance
|
When the system is busy, the compression component may be paged out of memory resulting in latency that can lead to warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using compression |
Trigger |
None |
Workaround |
Reduce compression workload |
|
7.7.1.7 |
Compression |
HU01614 |
All |
High Importance
|
After a node is upgraded hosts defined as TPGS may have paths set to inactive
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.6 or earlier with host type TPGS |
Trigger |
Upgrade to v7.7 or later |
Workaround |
None |
|
7.7.1.7 |
Hosts |
HU01636 |
V5000, V3700, V3500 |
High Importance
|
A connectivity issue with certain host SAS HBAs can prevent hosts from establishing stable communication with the storage controller
(show details)
Symptom |
Performance |
Environment |
Systems presenting storage to hosts with N2225 adapters |
Trigger |
Host with N2225 adapters running Windows Server 2012R2 |
Workaround |
None |
|
7.7.1.7 |
Hosts |
HU01638 |
All |
High Importance
|
When upgrading to v7.6 or later, if there is another cluster in the same zone which is at v5.1 or earlier then nodes will warmstart and the upgrade will fail
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.5 or earlier |
Trigger |
Upgrade to v7.6 or later with a cluster in the same zone running v6.1 or earlier |
Workaround |
Unzone any cluster running v5.1 or earlier from the cluster being upgraded |
|
7.7.1.7 |
System Update |
IT17564 |
All |
High Importance
|
All nodes in an I/O group may warmstart when a DRAID array experiences drive failures
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with DRAID arrays |
Trigger |
Failing drives |
Workaround |
None |
|
7.7.1.7 |
Distributed RAID |
IT19726 |
SVC |
High Importance
|
Warmstarts may occur when the attached SAN fabric is congested and HBA transmit paths become stalled preventing the HBA firmware from generating the completion for a FC command
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
SVC systems |
Trigger |
SAN congestion |
Workaround |
None |
|
7.7.1.7 |
Hosts |
IT21383 |
SVC, V7000, V5000 |
High Importance
|
Heavy I/O may provoke inconsistencies in resource allocation leading to node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems attached to slow backend controllers |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Reliability Availability Serviceability |
IT22376 |
V5000 |
High Importance
|
Upgrade of V5000 Gen 2 systems, with 16GB node canisters, can become stalled with multiple warmstarts on first node to be upgraded
(show details)
Symptom |
Loss of Redundancy |
Environment |
Storwize V5000 Gen 2 systems with 16GB memory in each node canister |
Trigger |
Upgrade to v7.6.1 or v7.7.1 |
Workaround |
None |
|
7.7.1.7 |
System Update |
HU00744 (reverted in 7.7.1.8) |
All |
Suggested
|
Single node warmstart due to an accounting issue within the cache component
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.3 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Cache |
HU00763 & HU01237 |
V7000, V5000, V3700, V3500 |
Suggested
|
A node warmstart may occur when a quorum disk is accessed at the same time as the login to that disk is closed
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize systems |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Quorum |
HU01098 |
All |
Suggested
|
Some older backend controller code levels do not support C2 commands resulting in 1370 entries in the Event Log for every detectmdisk
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later attached to backend controllers running older code levels |
Trigger |
Issue CLI command detectmdisk |
Workaround |
None |
|
7.7.1.7 |
Backend Storage |
HU01228 |
All |
Suggested
|
Automatic T3 recovery may fail due to the handling of quorum registration generating duplicate entries
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Reliability Availability Serviceability |
HU01229 |
V7000, V5000, V3700, V3500 |
Suggested
|
The DMP for a 3105 event does not identify the correct problem canister
(show details)
Symptom |
None |
Environment |
Storwize systems |
Trigger |
None |
Workaround |
Problem canister identified in Event log |
|
7.7.1.7 |
GUI Fix Procedure |
HU01332 |
All |
Suggested
|
Performance monitor and Spectrum Control show zero CPU utilisation for compression
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
System Monitoring |
HU01385 |
All |
Suggested
|
A warmstart may occur if a rmvolumecopy or rmrcrelationship command are issued on a volume while I/O is being forwarded to the associated copy
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6 or v7.7 that are using Hyperswap |
Trigger |
Issue an rmvolumecopy or rmrcrelationship command whilst hosts are still actively using the Hyperswap volume |
Workaround |
Do not remove a Hyperswap volume or relationship whilst hosts are still mapped to it |
|
7.7.1.7 |
HyperSwap |
HU01391 & HU01581 |
V7000, V5000, V3700, V3500 |
Suggested
|
Storwize systems may experience a warmstart due to an uncorrectable error in the SAS firmware
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize systems |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
Drives |
HU01430 |
V7000, V5000, V3700, V3500 |
Suggested
|
Memory resource shortages in systems with 8GB of RAM can lead to node warmstarts
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize Gen 1 systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
|
HU01457 |
V7000 |
Suggested
|
In a hybrid V7000 cluster where one I/O group supports 10k volumes and another does not some operations on volumes may incorrectly be denied in the GUI
(show details)
Symptom |
None |
Environment |
Systems running v7.7.1 or later |
Trigger |
None |
Workaround |
Performed required actions in the CLI |
|
7.7.1.7 |
Graphical User Interface |
HU01466 |
SVC, V7000, V5000 |
Suggested
|
Stretched cluster and HyperSwap I/O routing does not work properly due to incorrect ALUA data
(show details)
Symptom |
None |
Environment |
Systems running v7.8 or later using HyperSwap |
Trigger |
Move a host to an alternate I/O group |
Workaround |
Use chhost -site to set the correct site |
|
7.7.1.7 |
Hosts, HyperSwap |
HU01467 |
All |
Suggested
|
Failures in the handling of performance statistics files may lead to missing samples in Spectrum Control and other tools
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
Increase the sampling interval |
|
7.7.1.7 |
System Monitoring |
HU01469 |
V3700, V3500 |
Suggested
|
Resource exhaustion in the iSCSI component can result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize V3k systems using iSCSI |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
iSCSI |
HU01484 |
All |
Suggested
|
During a RAID array rebuild there may be node warmstarts
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
RAID |
HU01566 |
SVC |
Suggested
|
After upgrading, numerous 1370 errors are seen in the Event Log
(show details)
Symptom |
None |
Environment |
SVC CG8 systems containing old SSD drives (type Z16IZD2B-146) |
Trigger |
System Upgrade |
Workaround |
Ignore 1370 errors |
|
7.7.1.7 |
System Update |
HU01582 |
All |
Suggested
|
A compression issue in IP replication can result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7 or later using IP Replication |
Trigger |
None |
Workaround |
None |
|
7.7.1.7 |
IP Replication |
HU01474 |
SVC, V7000 |
HIPER
|
Host writes to a read-only secondary volume trigger I/O timeout warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
SVC SV1 & V7000 Gen 2+ systems running v7.7 or later using Remote Copy |
Trigger |
Host write I/O to a read-only secondary volume |
Workaround |
Do not attempt to issue host writes to a secondary volume while remote copy is active |
|
7.7.1.6 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01479 |
All |
HIPER
|
The handling of drive reseats can sometimes allow I/O to occur before the drive has been correctly failed resulting in offline MDisks
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.6 or later using DRAID |
Trigger |
Reseat a failed drive |
Workaround |
Rather than reseating the drive use the CLI or GUI to fail then unfail it |
|
7.7.1.6 |
Distributed RAID |
HU01483 |
All |
HIPER
|
mkdistributedarray command may get stuck in the prepare state. Any interaction with the volumes in that array will result in multiple warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1 or later |
Trigger |
mkdistributedarray command gets stuck in the prepare state; run rmarray or rmvdisk commands or generate I/O to volumes created on that array |
Workaround |
None |
|
7.7.1.6 |
Distributed RAID |
HU01500 |
All |
HIPER
|
Node warmstarts can occur when the iSCSI Ethernet MTU is changed
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1 with iSCSI connected hosts |
Trigger |
Change Ethernet MTU for iSCSI |
Workaround |
None |
|
7.7.1.6 |
iSCSI |
HU01225 & HU01330 & HU01412 |
All |
Critical
|
Node warmstarts due to inconsistencies arising from the way cache interacts with compression
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.3 or later with compressed volumes |
Trigger |
None |
Workaround |
None |
|
7.7.1.6 |
Cache, Compression |
HU01371 |
SVC, V7000, V5000 |
High Importance
|
A remote copy command related to HyperSwap may hang resulting in a warmstart of the config node
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.5 or later using HyperSwap |
Trigger |
None |
Workaround |
None |
|
7.7.1.6 |
HyperSwap |
HU01480 |
All |
High Importance
|
Under some circumstances the config node does not fail over properly when using IPv6 adversely affecting management access via GUI and CLI
(show details)
Symptom |
Configuration |
Environment |
Systems using IPv6 cluster addresses |
Trigger |
None |
Workaround |
None |
|
7.7.1.6 |
Command Line Interface, Graphical User Interface |
HU01473 |
All |
Suggested
|
EasyTier migrates an excessive number of cold extents to an overloaded nearline array
(show details)
Symptom |
Performance |
Environment |
Systems running v7.6.1 or later using EasyTier |
Trigger |
None |
Workaround |
None |
|
7.7.1.6 |
EasyTier |
HU01487 |
All |
Suggested
|
Small increase in read response time for source volumes with additional FlashCopy maps
(show details)
Symptom |
Performance |
Environment |
Systems running v7.5 or later using Global Mirror with Change Volumes |
Trigger |
Add FlashCopy mappings to GMCV source volumes |
Workaround |
Avoid using incremental FlashCopy |
|
7.7.1.6 |
FlashCopy, Global Mirror with Change Volumes |
HU01498 |
All |
Suggested
|
GUI may be exposed to CVE-2017-5638 (see Section 3.1)
|
7.7.1.6 |
Security |
IT18752 |
All |
Suggested
|
When the config node processes an lsdependentvdisks command, issued via the GUI, that has a large number of objects in its parameters, it may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7 or later |
Trigger |
Issue lsdependentvdisks command, via the GUI, for a large number of objects |
Workaround |
Issue those commands, via the CLI |
|
7.7.1.6 |
Graphical User Interface |
HU01193 |
All |
HIPER
|
A drive failure whilst an array rebuild is in progress can lead to both nodes in an I/O group warmstarting
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
Drive failure |
Workaround |
None |
|
7.7.1.5 |
Distributed RAID |
HU01382 |
All |
HIPER
|
Mishandling of extent migration following a rmarray command can lead to multiple simultaneous node warmstarts with a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1 or later using DRAID |
Trigger |
Executing a rmarray command |
Workaround |
Avoid using rmarray command |
|
7.7.1.5 |
Distributed RAID |
HU01340 |
All |
Critical
|
A port translation issue between v7.5 or earlier and v7.7.0 or later requires a Tier 2 recovery to complete an upgrade
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.5 or earlier |
Trigger |
Upgrade to v7.7.x |
Workaround |
None |
|
7.7.1.5 |
System Update |
HU01392 |
All |
Critical
|
Under certain rare conditions FC mappings not in a consistency group can be added to a special internal consistency group resulting in a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
FlashCopy |
HU01410 |
SVC |
Critical
|
An issue in the handling of FlashCopy map preparation can cause both nodes in an I/O group to be put into service state
(show details)
Symptom |
Loss of Access to Data |
Environment |
SVC systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
FlashCopy |
HU01223 |
All |
High Importance
|
The handling of a rebooted nodes return to the cluster can occasionally become delayed resulting in a stoppage of inter-cluster relationships
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.3 or later using MetroMirror |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
Metro Mirror |
HU01254 |
SVC |
High Importance
|
A fluctuation of input AC power can cause a 584 error on a node
(show details)
Symptom |
Loss of Redundancy |
Environment |
SVC CF8/CG8 systems |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
Reliability Availability Serviceability |
HU01402 |
V7000 |
High Importance
|
Nodes can power down unexpectedly as they are unable to determine from their partner whether power is available
(show details)
Symptom |
Loss of Redundancy |
Environment |
V7000 Gen 1 systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
Reliability Availability Serviceability |
HU01409 |
All |
High Importance
|
Cisco Nexus 3000 switches at v5.0(3) have a defect which prevents a config node IP address changing in the event of a fail over
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems connected to Cisco Nexus 3000 switches |
Trigger |
Config node fail over |
Workaround |
None |
|
7.7.1.5 |
Reliability Availability Serviceability |
IT14917 |
All |
High Importance
|
Node warmstarts due to a timing window in the cache component. For more details refer to this Flash
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.4 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
Cache |
HU00831 |
All |
Suggested
|
Single node warmstart due to hung I/O caused by cache deadlock
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.2 or later |
Trigger |
Hosts sending many large block IOs that use many credits per I/O |
Workaround |
None |
|
7.7.1.5 |
Cache |
HU01022 |
SVC, V7000 |
Suggested
|
Fibre channel adapter encountered a bit parity error resulting in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6 or later using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
Hosts |
HU01269 |
All |
Suggested
|
A rare timing conflict between two process may lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
|
HU01399 |
All |
Suggested
|
For certain config nodes the CLI Help commands may not work
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
Use Knowledge Center |
|
7.7.1.5 |
Command Line Interface |
HU01432 |
All |
Suggested
|
Node warmstart due to an accounting issue within the cache component
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.3 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
Cache |
IT17302 |
V5000, V3700, V3500 |
Suggested
|
Unexpected 45034 1042 entries in the Event Log
(show details)
Symptom |
None |
Environment |
Systems running v7.7 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.5 |
System Monitoring |
IT18086 |
All |
Suggested
|
When a volume is moved between I/O groups a node may warmstart
(show details)
Symptom |
None |
Environment |
Systems running v7.7.1 or later |
Trigger |
No warmstarts since upgrade. Move a volume between I/O groups |
Workaround |
None |
|
7.7.1.5 |
|
HU01379 |
All |
HIPER
|
Resource leak in the handling of Read Intensive drives leads to offline volumes
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v7.7.1 or later using read intensive drives |
Trigger |
None |
Workaround |
None |
|
7.7.1.4 |
Drives |
HU01783 |
All |
Critical
|
Replacing a failed drive in a DRAID array, with a smaller drive, may result in multiple Tier 2 recoveries putting all nodes in service state with error 564 and/or 550
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
Replacing a drive with one of less capacity |
Workaround |
Ensure replacement drives are the same capacity |
|
7.7.1.4 |
Distributed RAID |
HU01347 |
All |
High Importance
|
During an upgrade to v7.7.1 a deadlock in node communications can occur leading to a timeout and node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.7.0 or earlier using SE volumes |
Trigger |
Upgrade to v7.7.1 |
Workaround |
None |
|
7.7.1.4 |
Thin Provisioning |
HU01381 |
All |
High Importance
|
A rare timing issue in FlashCopy may lead to a node warmstarting repeatedly and then entering a service state
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.2 or later using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.7.1.4 |
FlashCopy |
HU01247 |
All |
Suggested
|
When a FlashCopy consistency group is stopped more than once in rapid succession a node warmstart may result
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using FlashCopy |
Trigger |
Stop the same FlashCopy consistency group twice in rapid succession using -force on second attempt |
Workaround |
Avoid stopping the same flash copy consistency group more than once |
|
7.7.1.4 |
FlashCopy |
HU01323 |
All |
Suggested
|
Systems using Volume Mirroring that upgrade to v7.7.1.x and have a storage pool go offline may experience a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7.1 or later using volume mirroring |
Trigger |
1. Mirrored volumes are created at a pre-7.7.1 code level; 2. The system is upgraded to 7.7.1.1 or later; 3. A storage pool goes offline before the nodes next restart. |
Workaround |
None |
|
7.7.1.4 |
Volume Mirroring |
HU01374 |
All |
Suggested
|
Where an issue with Global Mirror causes excessive I/O delay, a timeout may not function resulting in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7.0 or later using Global Mirror |
Trigger |
None |
Workaround |
None |
|
7.7.1.4 |
Global Mirror |
HU01226 |
All |
High Importance
|
Changing max replication delay from the default to a small non-zero number can cause hung IOs leading to multiple node warmstarts and a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.6 or later using Global Mirror |
Trigger |
Changing max replication delay to a small non-zero number |
Workaround |
Do not change max replication delay to below 30 seconds |
|
7.7.1.3 |
Global Mirror |
HU01257 |
All |
High Importance
|
Large (>1MB) write IOs to volumes can lead to a hung I/O condition resulting in node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.7.1 or later |
Trigger |
Large (>1MB) write IOs |
Workaround |
Try to keep maximum I/O write transfer size to volumes below 1MB |
|
7.7.1.3 |
|
HU01386 |
All |
High Importance
|
Where latency between sites is greater than 1ms host write latency can be adversely impacted. This is can be more likely in the presence of large I/O transfer sizes or high IOPS
(show details)
Symptom |
Performance |
Environment |
Systems running v7.5 or later using HyperSwap with no Remote Copy partnerships |
Trigger |
Inter-site latency is greater than 1ms |
Workaround |
Keep inter-site latency below 1ms. Use Metro Mirror or Global Mirror replication |
|
7.7.1.3 |
HyperSwap |
HU01017 |
All |
Suggested
|
The result of CLI commands are sometimes not promptly presented in the GUI
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
In the GUI, navigate to "GUI Preferences" ("GUI Preferences->General" in v7.6 or later) and refresh the GUI cache. |
|
7.7.1.3 |
Graphical User Interface |
HU01227 |
All |
Suggested
|
High volumes of events may cause the email notifications to become stalled
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later |
Trigger |
More than 15 events per second |
Workaround |
None |
|
7.7.1.3 |
System Monitoring |
HU01234 |
All |
Suggested
|
After upgrade to 7.6 or later iSCSI hosts may incorrectly be shown as offline in the CLI
(show details)
Symptom |
None |
Environment |
Systems with iSCSI connected hosts |
Trigger |
Upgrade to v7.6 or later from v7.5 or earlier |
Workaround |
None |
|
7.7.1.3 |
iSCSI |
HU01251 |
V7000, V5000, V3700, V3500 |
Suggested
|
When following the DMP for a 1685 event, if the option for drive reseat has already been attempted is selected, the process to replace a drive is not started
(show details)
Symptom |
None |
Environment |
Storwize systems running v7.3 or later |
Trigger |
DMP for a 1685 event is run. Select "drive reseat has already been attempted" option |
Workaround |
Manually replace drive |
|
7.7.1.3 |
GUI Fix Procedure |
HU01292 |
All |
Suggested
|
Under some circumstances the re-calculation of grains to clean can take too long after a FlashCopy done event has been sent resulting in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7.0 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.3 |
FlashCopy |
IT17102 |
All |
Suggested
|
Where the maximum number of I/O requests for a FC port has been exceeded, if a SCSI command, with an unsupported opcode, is received from a host then the node may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.5 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.3 |
|
HU01272 |
All |
HIPER
|
Replacing a drive in a system with a DRAID array can result in T2 recovery warmstarts. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1.0 or later using DRAID |
Trigger |
Replacing a drive |
Workaround |
Change the failed drive to unused prior to removal by using "chdrive -use unused $driveid" before it is removed from the system; Physically replace the drive; Manually add the drive to the array using "charraymember -newdrive $driveid"; Remove the ghost drive by using "chdrive -use unused $ghostdriveid". |
|
7.7.1.2 |
Distributed RAID |
HU01118 |
V7000, V5000, V3700, V3500 |
HIPER
|
Due to a firmware issue both nodes in a V7000 Gen 2 may be powered off
(show details)
Symptom |
Loss of Access to Data |
Environment |
Storwize V3500, V3700, V5000 & V7000 Gen 2 systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
|
HU01208 |
All |
HIPER
|
After upgrading to v7.7 or later from v7.5 or earlier and then creating a DRAID array, with a node reset, the system may encounter repeated node warmstarts which will require a Tier 3 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.5.0.0 or earlier |
Trigger |
Code upgrade, create DRAID array and then reset nodes |
Workaround |
None |
|
7.7.1.1 |
Distributed RAID |
HU00271 |
All |
High Importance
|
An extremely rare timing window condition in the way GM handles write sequencing may cause multiple node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using GM |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
Global Mirror |
HU00734 |
All |
High Importance
|
Multiple node warmstarts due to deadlock condition during RAID group rebuild
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All |
Trigger |
Array rebuilds |
Workaround |
Hold one node in service state till the array rebuild finishes |
|
7.7.1.1 |
|
HU01109 |
SVC, V7000, V5000 |
High Importance
|
Multiple nodes can experience a lease expiry when a FC port is having communications issues
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.6 or later using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
Reliability Availability Serviceability |
HU01140 |
All |
High Importance
|
EasyTier may unbalance the workloads on MDisks using specific Nearline SAS drives due to incorrect thresholds for their performance
(show details)
Symptom |
Performance |
Environment |
Systems running v7.3 or later using EasyTier |
Trigger |
None |
Workaround |
Add Enterprise-class drives to the MDisk or MDisk group that is experiencing unbalanced workloads |
|
7.7.1.1 |
EasyTier |
HU01141 |
All |
High Importance
|
Node warmstart (possibly due to a network problem) when a CLI mkippartnership is issued. This may lead to loss of the config node requiring a Tier 2 recovery
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.5 or later using IP Replication |
Trigger |
Enter mkippartnership CLI command |
Workaround |
Ensure partner cluster IP can be pinged before issuing a mkippartnership CLI command |
|
7.7.1.1 |
IP Replication |
HU01182 |
SVC, V7000, V5000 |
High Importance
|
Node warmstarts due to 16Gb HBA firmware receiving an invalid SCSI TUR command
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.4 or later using 16Gb HBAs |
Trigger |
16G FC HBA receives a SCSI TUR command with Total XFER LEN > 0 |
Workaround |
None |
|
7.7.1.1 |
|
HU01184 |
All |
High Importance
|
When removing multiple MDisks node warmstarts may occur
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.6 or later |
Trigger |
Multiple rmmdisk commands issued in rapid succession |
Workaround |
Remove MDisks one at a time and let migration complete before proceeding to next MDisk removal |
|
7.7.1.1 |
|
HU01185 |
All |
High Importance
|
iSCSI target closes connection when there is a mismatch in sequence number
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with iSCSI connected hosts |
Trigger |
Out of sequence I/O |
Workaround |
None |
|
7.7.1.1 |
iSCSI |
HU01210 |
SVC |
High Importance
|
A small number of systems have broken, or disabled, TPMs. For these systems the generation of a new master key may fail preventing the system joining a cluster
(show details)
Symptom |
Loss of Redundancy |
Environment |
CG8 systems running v7.6.1 or later |
Trigger |
Broken or disabled TPM |
Workaround |
None |
|
7.7.1.1 |
|
HU01221 |
SVC, V7000, V5000 |
High Importance
|
Node warmstarts due to an issue with the state machine transition in 16Gb HBA firmware
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.6 or later using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
|
HU01250 |
All |
High Importance
|
When using lsvdisklba to find a bad block on a compressed volume, the volume can go offline
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v7.3 or later that are using compressed volumes |
Trigger |
Using lsvdisklba |
Workaround |
None |
|
7.7.1.1 |
Compression |
HU01516 |
All |
High Importance
|
When node configuration data exceeds 8K in size some user defined settings may not be stored permanently resulting in node warmstarts
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with large, complex configurations |
Trigger |
A high number changes to node configuration |
Workaround |
Minimise node configuration changes |
|
7.7.1.1 |
Reliability Availability Serviceability |
IT16148 |
All |
High Importance
|
When accelerate mode is enabled due to the way promote/swap plans are prioritized over demote EasyTier is only demoting 1 extent every 5 minutes
(show details)
Symptom |
Performance |
Environment |
Systems running v7.5 or later using EasyTier |
Trigger |
Over use of accelerate mode |
Workaround |
Only use accelerate mode when migrating data between pools |
|
7.7.1.1 |
EasyTier |
IT16337 |
SVC, V7000, V5000 |
High Importance
|
Hardware offloading in 16G FC adapters has introduced a deadlock condition that causes many driver commands to time out leading to a node warmstart. For more details refer to this Flash
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.4 or later using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
|
HU01024 |
V7000, V5000, V3700, V3500 |
Suggested
|
A single node warmstart may occur when the SAS firmwares ECC checking detects a single bit error. The warmstart clears the error condition in the SAS chip
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.4 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
|
HU01050 |
All |
Suggested
|
DRAID rebuild incorrectly reports event code 988300
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
Distributed RAID |
HU01063 |
SVC, V7000, V5000 |
Suggested
|
3PAR controllers do not support OTUR commands resulting in device port exclusions
(show details)
Symptom |
None |
Environment |
Systems virtualising 3PAR storage |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
Backend Storage |
HU01155 |
All |
Suggested
|
When a lsvdisklba or lsmdisklba command is invoked, for an MDisk with a back end issue, a node warmstart may occur
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7.0 or earlier using compression |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
Compression |
HU01187 |
All |
Suggested
|
Circumstances can arise where more than one array rebuild operation can share the same CPU core resulting in extended completion times
(show details)
Symptom |
Performance |
Environment |
Systems running v7.4 or later |
Trigger |
None |
Workaround |
Avoid R5 array configurations |
|
7.7.1.1 |
|
HU01194 |
All |
Suggested
|
A single node warmstart may occur if CLI commands are received from the VASA provider in very rapid succession. This is caused by a deadlock condition which prevents the subsequent CLI command from completing
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6 or later that are using vVols |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
vVols |
HU01198 |
All |
Suggested
|
Running the Comprestimator svctask analyzevdiskbysystem command may cause the config node to warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6 or later using Comprestimator |
Trigger |
Run svctask analyzevdiskbysystem |
Workaround |
Avoid using svctask analyzevdiskbysystem |
|
7.7.1.1 |
Comprestimator |
HU01214 |
All |
Suggested
|
GUI and snap missing EasyTier heatmap information
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later |
Trigger |
None |
Workaround |
Download the files individually via CLI |
|
7.7.1.1 |
Support Data Collection |
HU01219 |
SVC, V7000, V5000 |
Suggested
|
Single node warmstart due to an issue in the handling of ECC errors within 16G HBA firmware
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6 or later using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
|
HU01244 |
All |
Suggested
|
When a node is transitioning from offline to online it is possible for excessive CPU time to be used on another node in the cluster which may lead to a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.3 or later |
Trigger |
None |
Workaround |
None |
|
7.7.1.1 |
|
HU01258 |
SVC |
Suggested
|
A compressed volume copy will result in an unexpected 1862 message when site/node fails over in a stretched cluster configuration
(show details)
Symptom |
None |
Environment |
SVC systems running v7.4 or later in a stretched cluster configuration |
Trigger |
Site/node failover |
Workaround |
None |
|
7.7.1.1 |
Compression |