HU02342 |
All |
Critical
|
Occasionally when an offline drive returns to online state later than its peers in the same RAID array there can be multiple node warmstarts that send nodes into a service state
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.15 |
RAID |
HU02406 |
All |
Critical
|
An interoperability issue between Cisco NX-OS firmware and the Spectrum Virtualize Fibre Channel driver can cause a node warmstart on NPIV failback (for example during an upgrade) with the potential for a loss of access. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using NPIV that are connected to Cisco SAN equipment running NX-OS 8.4(2c) or later |
Trigger |
Initiate an NPIV failback operation by, for example, performing an upgrade |
Workaround |
Disable NPIV (which will require any hot spare nodes to be removed first) |
|
7.8.1.15 |
Interoperability |
HU02471 |
All |
Critical
|
After starting a FlashCopy map with -restore in a graph with a GMCV secondary disk that was stopped with -access there can be a data integrity issue
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems using GMCV |
Trigger |
Given a configuration where there is a GMCV secondary volume A that has a cascade of two FlashCopy maps to volumes B and C (A -> B -> C) The mapping B -> C has been started and GMCV is running Stop the GMCV relationship with -access Start FlashCopy map A -> B with -restore Any I/O to volume A will corrupt data on volume C |
Workaround |
Wait for a GMCV backward map to complete before starting related FlashCopy maps with -restore |
|
7.8.1.15 |
FlashCopy, Global Mirror with Change Volumes |
HU02332 & HU02336 |
All |
Suggested
|
When an I/O is received, from a host, with invalid or inconsistent SCSI data but a good checksum it may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.15 |
Hosts |
HU02429 |
All |
Critical
|
System can go offline shortly after changing the SMTP settings using the chemailserver command via the GUI
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8.1 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.14 |
System Monitoring |
HU02277 |
All |
HIPER
|
RAID parity scrubbing can become stalled causing an accumulation of media errors leading to multiple drive failures with the possibility of data integrity loss. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems with model MZILS3T8HMLH read intensive SSDs at drive firmware MS24 are particularly susceptible to the data integrity (DI) issue. Other drive types may see multiple failures without DI issue |
Trigger |
None |
Workaround |
None |
|
7.8.1.13 |
RAID |
HU02338 |
All |
HIPER
|
An issue in the setting up of reverse FlashCopy mappings can cause the background copy to finish prematurely providing an incomplete target image
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.8.1.13 |
FlashCopy |
HU02222 |
All |
Critical
|
Where the source volume of an incremental FlashCopy map is also a Metro or Global Mirror target volume that is using a change volume or is a Hyperswap volume, then there is a possibility that not all data will be copied to the FlashCopy target. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems using Remote Copy |
Trigger |
None |
Workaround |
None |
|
7.8.1.13 |
Global Mirror with Change Volumes |
HU02201 & HU02221 |
All |
High Importance
|
Shortly after upgrading drive firmware, specific drive models can fail due to Too many long IOs to drive for too long errors
(show details)
Symptom |
Degraded Operation |
Environment |
Systems with the following drive models:
- ST300MM0009 (300GB) - B5B8
- ST600MM0009 (600GB) - B5B8
- ST900MM0009 (900GB) - B5B8
- ST1200MM0009 (1200GB) - B5B8
- ST1200MM0129 (1800GB) - B5C9
- ST2400MM0129 (2400GB) - B5C9
- ST300MP0006 (300GB) - B6AA
- ST600MP0006 (600GB) - B6AA
- ST900MP0146 (900GB) - B6CB
|
Trigger |
None |
Workaround |
None |
|
7.8.1.13 |
Drives |
HU02238 |
All |
HIPER
|
Force-stopping a FlashCopy map, where the source volume is a Metro or Global Mirror target volume, may cause other FlashCopy maps to return invalid data if they are not 100% copied, in specific configurations. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.8.1.12 |
FlashCopy, Global Mirror, Metro Mirror |
HU01970 |
All |
Critical
|
When a GMCV relationship is stopped, with the -access option, and the secondary volume is immediately deleted with -force, then all nodes may repeatedly warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using GMCV |
Trigger |
Stop a GMCV relationship with -access and immediately delete the secondary volume |
Workaround |
Do not remove secondary volume, with -force, if the backward FC map from the secondary change volume to the secondary volume is still in progress |
|
7.8.1.12 |
Global Mirror with Change Volumes |
HU02197 |
All |
Critical
|
Bulk volume removals can adversely impact related FlashCopy mappings leading to a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1, or later, using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.8.1.12 |
FlashCopy |
IT25367 |
All |
Critical
|
A T2 recovery may occur when an attempt is made to upgrade, or downgrade, the firmware for an unsupported drive type
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Attempt to upgrade/downgrade the firmware for an unsupported drive type |
Workaround |
None |
|
7.8.1.12 |
Drives |
HU01832 |
All |
Suggested
|
Creation and distribution of the config file may cause an out-of-memory condition, leading to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.12 |
Reliability Availability Serviceability |
HU01868 |
All |
Suggested
|
After deleting an encrypted external MDisk, it is possible for the encrypted status of volumes to change to no, even though all remaining MDisks are encrypted
(show details)
Symptom |
None |
Environment |
Systems using encryption |
Trigger |
Delete an encrypted external MDisk |
Workaround |
Ensure that all MDisks in the MDisk group are encrypted - this will ensure that data is encrypted |
|
7.8.1.12 |
Encryption |
HU01917 |
All |
Suggested
|
Chrome browser support requires a self-signed certificate to include subject alternate name
(show details)
Symptom |
None |
Environment |
Systems accessed using the Chrome browser |
Trigger |
None |
Workaround |
Accept invalid certificate |
|
7.8.1.12 |
Graphical User Interface |
HU02014 |
SVC |
HIPER
|
After a loss of power, where a node has a dead CMOS battery, it will fail to restart correctly. It is possible for both nodes in an I/O group to experience this issue
(show details)
Symptom |
Loss of Access to Data |
Environment |
SVC systems using SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Reliability Availability Serviceability |
HU01887 |
All |
Critical
|
In circumstances where host configuration data becomes inconsistent, across nodes, an issue in the CLI policing code may cause multiple warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Host Clusters |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Command Line Interface, Host Cluster |
HU02043 |
All |
Critical
|
Collecting a snap can cause nodes to run out of boot drive space and go offline with node error 565
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Taking many snap data collections on the same config node |
Workaround |
Manually delete unneeded snaps from the boot drive |
|
7.8.1.11 |
Support Data Collection |
HU02063 |
All |
Critical
|
HyperSwap clusters with only two surviving nodes may experience warmstarts on both of those nodes where rcbuffersize is set to 512MB
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Hyperswap |
Trigger |
None |
Workaround |
Reduce rcbuffersize to a value less than 512 |
|
7.8.1.11 |
HyperSwap |
IT26257 |
All |
Critical
|
Starting a relationship, when the remote volume is offline, may result in a T2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Hyperswap |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
HyperSwap |
HU01836 |
All |
High Importance
|
When an auxiliary volume is moved an issue with pausing the master volume can lead to node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using Hyperswap |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
HyperSwap |
HU01923 |
All |
High Importance
|
An issue in the way Global Mirror handles write sequence numbers >512 may cause multiple node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using GM |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Global Mirror |
HU01944 |
All |
High Importance
|
Proactive host failover not waiting for 25 seconds before allowing nodes to go offline during upgrades or maintenance
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Reliability Availability Serviceability |
HU01952 |
SVC, V7000 |
High Importance
|
When the compression accelerator hardware driver detects an uncorrectable error the node will reboot
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems using compression cards |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Compression |
HU02049 |
All |
High Importance
|
GUI session handling has an issue that can generate many exceptions, adversely impacting GUI performance
(show details)
Symptom |
Performance |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Graphical User Interface |
HU02149 |
SVC |
High Importance
|
When an Enhanced Stretch Cluster is using NPIV, in transitional mode, the path priority is not being reported correctly to some hosts
(show details)
Symptom |
Performance |
Environment |
Systems in an Enhanced Stretch Cluster topology that are using NPIV in its transitional mode |
Trigger |
None |
Workaround |
Manually set the preferred path if possible within the host's MPIO settings. Run NPIV in enabled or disabled mode |
|
7.8.1.11 |
Hosts |
HU01830 |
All |
Suggested
|
Missing security-enhancing HTTP response headers
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Security |
HU01863 |
All |
Suggested
|
In rare circumstances, a drive replacement may result in a ghost drive (i.e. a drive with the same ID as the replaced drive stuck in a permanently offline state)
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
Drive replacement |
Workaround |
None |
|
7.8.1.11 |
Drives |
HU01892 |
All |
Suggested
|
LUNs of greater than 2TB, presented by HP XP7 storage controllers, are not supported
(show details)
Symptom |
Configuration |
Environment |
Systems with HP XP7 backend controllers |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Backend Storage |
HU01981 |
All |
Suggested
|
Although an issue, in the HBA firmware, is handled correctly it can still cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Reliability Availability Serviceability |
HU01988 |
All |
Suggested
|
In the Monitoring -> 3D view page, the "export to csv" button does not function
(show details)
Symptom |
None |
Environment |
Systems running v8.1 or earlier |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Graphical User Interface |
HU02085 |
All |
Suggested
|
Freeze time of Global Mirror remote copy consistency groups may not be updated correctly in certain scenarios
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later using Global Mirror |
Trigger |
None |
Workaround |
None |
|
7.8.1.11 |
Global Mirror |
HU02102 |
All |
Suggested
|
Excessive processing time required for FlashCopy bitmap operations, associated with large (> 20TB) Global Mirror change volumes, may lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Global Mirror with Change Volumes where some GMCV volumes are >20TB |
Trigger |
None |
Workaround |
Limit GMCV volume capacity to 20TB or less |
|
7.8.1.11 |
Global Mirror with Change Volumes |
HU02126 |
SVC, V7000, V5000 |
Suggested
|
There is a low probability that excessive SSH connections may trigger a single node warmstart on the configuration node
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with Gen 1 and 2 hardware |
Trigger |
More than one SSH connection attempt per second might occasionally cause the config node to warmstart |
Workaround |
Reduce frequency of SSH connections |
|
7.8.1.11 |
Command Line Interface |
HU01781 |
All |
Critical
|
An issue with workload balancing in the kernel scheduler can deprive some processes of the necessary resource to complete successfully resulting in a node warmstarts, that may impact performance, with the possibility of a loss of access to volumes
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.10 |
|
HU01888 & HU01997 |
All |
Critical
|
An issue with restore mappings, in the FlashCopy component, can cause an I/O group to warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.10 |
FlashCopy |
HU01972 |
All |
High Importance
|
When an array is in a quiescing state, for example where a member has been deleted, I/O may become pended leading to multiple warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All systems |
Trigger |
Delete an array member using "charraymember -used unused" command |
Workaround |
None |
|
7.8.1.10 |
Distributed RAID, RAID |
HU00744 |
All |
Suggested
|
Single node warmstart due to an accounting issue within the cache component
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.3 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.10 |
Cache |
HU00921 |
All |
Suggested
|
A node warmstart may occur when an MDisk state change gives rise to duplicate discovery processes
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.10 |
|
HU01737 |
All |
Suggested
|
On the Update System screen, for Test Only, if a valid code image is selected, in the Run Update Test Utility dialog, then clicking the Test button will initiate a system update
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
Select a valid code image in the "Run Update Test Utility" dialog and click "Test" button |
Workaround |
Do not select a valid code image in the "Test utility" field of the "Run Update Test Utility" dialog |
|
7.8.1.10 |
System Update |
HU01915 & IT28654 |
All |
Suggested
|
Systems, with encryption enabled, that are using key servers to manage encryption keys, may fail to connect to the key servers if the servers SSL certificates are part of a chain of trust
(show details)
Symptom |
None |
Environment |
Systems with encryption enabled |
Trigger |
None |
Workaround |
None |
|
7.8.1.10 |
Encryption |
HU01617 |
All |
HIPER
|
Due to a timing window issue, stopping a FlashCopy mapping, with the -autodelete option, may result in a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
Avoid stopping FlashCopy mappings with the -autodelete option |
|
7.8.1.9 |
FlashCopy |
HU01708 |
All |
HIPER
|
A node removal operation during an array rebuild can cause a loss of parity data leading to bad blocks
(show details)
Symptom |
Data Integrity Loss |
Environment |
All systems |
Trigger |
Removing a node during an array rebuild |
Workaround |
Do not remove nodes during an array rebuild |
|
7.8.1.9 |
RAID |
HU01865 |
All |
HIPER
|
When creating a HyperSwap relationship, using addvolumecopy (or similar methods), the system should perform a synchronisation operation to copy the data from the original copy to the new copy. In some rare cases this synchronisation is skipped, leaving the new copy with bad data (all zeros)
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems running v7.5 or later using HyperSwap |
Trigger |
None |
Workaround |
None |
|
7.8.1.9 |
HyperSwap |
HU01913 |
All |
HIPER
|
A timing window issue in the DRAID6 rebuild process can cause node warmstarts with the possibility of a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
7.8.1.9 |
Distributed RAID |
HU01723 |
All |
Critical
|
A timing window issue, around nodes leaving and re-joining clusters, can lead to hung I/O and node warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.9 |
Reliability Availability Serviceability |
HU01876 |
All |
Critical
|
Where systems are connected to controllers, that have FC ports that are capable of acting as initiators and targets, when NPIV is enabled then node warmstarts can occur
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems, with NPIV enabled, attached to host ports that can act as SCSI initiators and targets |
Trigger |
Zone host initiator and target ports in with the target port WWPN then enable NPIV |
Workaround |
Unzone host or disable NPIV |
|
7.8.1.9 |
Backend Storage |
IT27460 |
All |
Critical
|
Lease expiry can occur between local nodes when remote connection is lost, due to the mishandling of messaging credits
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
Use four ports for local to local node communications, on at least two separate fibre channel adapters per node. Port mask so that all four are usable. Use a different fibre channel adapter than the above two adapters for remote port communications. If there are issues with the FCIP tunnel, temporarily block that until it is fixed. |
|
7.8.1.9 |
Reliability Availability Serviceability |
IT29040 |
All |
Critical
|
Occasionally a DRAID rebuild, with drives of 8TB or more, can encounter an issue which causes node warmstarts and potential loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID with drives of 8TB or more |
Trigger |
None |
Workaround |
None |
|
7.8.1.9 |
Distributed RAID, RAID |
HU01907 |
SVC |
High Importance
|
An issue in the handling of the power cable sense registers can cause a node to be put into service state with a 560 error
(show details)
Symptom |
Loss of Redundancy |
Environment |
SVC systems using SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
7.8.1.9 |
Reliability Availability Serviceability |
HU01485 |
SVC |
Suggested
|
When a SV1 node is started, with only one PSU powered, powering up the other PSU will not extinguish the Power Fault LED.Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed
(show details)
Symptom |
None |
Environment |
SVC systems using SV1 model nodes |
Trigger |
Power up node with only one PSU powered. Power Fault LED is lit. Power up other PSU. Power Fault LED remains lit. |
Workaround |
Ensure both PSUs are powered before starting node |
|
7.8.1.9 |
System Monitoring |
HU01659 |
SVC |
Suggested
|
Node Fault LED can be seen to flash in the absence of an error condition.Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed
(show details)
Symptom |
None |
Environment |
SVC systems using SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
7.8.1.9 |
System Monitoring |
HU01849 |
All |
Suggested
|
An excessive number of SSH sessions may lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
Initiate a large number of SSH sessions (e.g. one session every 5 seconds) |
Workaround |
Avoid initiating excessive numbers of SSH sessions |
|
7.8.1.9 |
System Monitoring |
IT26049 |
All |
Suggested
|
An issue with CPU scheduling may cause the GUI to respond slowly
(show details)
Symptom |
None |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.9 |
Graphical User Interface |
HU01492 & HU02024 |
SVC, V7000, V5000 |
HIPER
|
All ports of a 16Gb HBA can be affected when a single port is congested. This can lead to lease expiries if all ports used for inter-node communication are on the same FC adapter
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using 16Gb HBAs |
Trigger |
All ports used for inter-node communication are on the same FC adapter and a port on that adapter experiences congestion |
Workaround |
Separate inter-node traffic so that multiple adapters are used |
|
7.8.1.8 |
Reliability Availability Serviceability |
HU01726 |
All |
HIPER
|
A slow raid member drive in an MDisk may cause node warmstarts and the MDisk to go offline for a short time
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using DRAID |
Trigger |
Failing drives |
Workaround |
None |
|
7.8.1.8 |
Distributed RAID |
HU01940 |
All |
HIPER
|
Changing the use of a drive can cause a Tier 2 recovery (warmstarts on all nodes in the cluster). This occurs only if the drive change occurs within a small timing window, so the probability of the issue occurring is low
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Change of drive use |
Workaround |
None |
|
7.8.1.8 |
Drives |
HU01572 |
All |
Critical
|
SCSI 3 commands from unconfigured WWPNs may result in multiple warmstarts leading to a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with iSCSI connected hosts |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
iSCSI |
HU01678 |
All |
Critical
|
Entering an invalid parameter in the addvdiskaccess command may initiate a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Include an invalid argument in an addvdiskaccess command |
Workaround |
Use care when entering addvdiskaccess CLI commands |
|
7.8.1.8 |
Command Line Interface |
HU01735 |
All |
Critical
|
Multiple power failures can cause a RAID array to get into a stuck state leading to offline volumes
(show details)
Symptom |
Offline Volumes |
Environment |
All systems |
Trigger |
Multiple power failures |
Workaround |
None |
|
7.8.1.8 |
RAID |
HU01774 |
All |
Critical
|
After a failed mkhost command for an iSCSI host any I/O from that host will cause multiple warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with iSCSI connected hosts |
Trigger |
mkhost command fails for an iSCSI host |
Workaround |
For iSCSI hosts prevent I/O from the host if the mkhost operation fails |
|
7.8.1.8 |
iSCSI |
HU01799 |
All |
Critical
|
Timing window issue can affect operation of the HyperSwap addvolumecopy command causing all nodes to warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using HyperSwap |
Trigger |
None |
Workaround |
Ensuring that the volume is not in a FlashCopy mapping will allow the use of the addvolumecopy. Alternatively the addvolumecopy command combines a number of CLI commands for convenience. Use these CLI commands individually to achieve the same outcome |
|
7.8.1.8 |
HyperSwap |
HU01825 |
All |
Critical
|
Invoking a chrcrelationship command when one of the relationships in a consistency group is running in the opposite direction to the others may cause a node warmstart followed by a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using HyperSwap |
Trigger |
A relationship running in one direction is added to a consistency group running in the other direction whilst one of the FlashCopy maps associated with the HyperSwap relationship is still stopping/cleaning |
Workaround |
Do not add a relationship to a consistency group if they are running in opposite directions (i.e. the Primary of the consistency group and the Primary of the relationship are on different sites); Do not add a relationship to a consistency group if the relationship still has one of its FlashCopy maps in the stopping state. The clean progress needs to reach 100 percent before the relationship can be safely added. |
|
7.8.1.8 |
FlashCopy |
HU01847 |
All |
Critical
|
FlashCopy handling of medium errors across a number of drives on backend controllers may lead to multiple node warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8.1 or later using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
FlashCopy |
HU01899 |
All |
Critical
|
In a HyperSwap cluster, when the primary I/O group has a dead domain, nodes will repeatedly warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using HyperSwap |
Trigger |
A primary volume is offline and remote copy relationship is in a quiesced state |
Workaround |
None |
|
7.8.1.8 |
HyperSwap |
IT25850 |
All |
Critical
|
I/O performance may be adversely affected towards the end of DRAID rebuilds. For some systems there may be multiple warmstarts leading to a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Distributed RAID |
HU01507 |
All |
High Importance
|
Until the initial synchronisation process completes, high system latency may be experienced when a volume is created with two compressed copies or when space-efficient copy is added to a volume with an existing compressed copy
(show details)
Symptom |
Performance |
Environment |
All systems |
Trigger |
Create a volume with two compressed copies or add a space-efficient copy to a volume with an existing compressed copy |
Workaround |
Avoid: creating a new volume with two compressed copies; adding a SE volume copy to a volume that already possesses a compressed copy |
|
7.8.1.8 |
Volume Mirroring |
HU01579 |
All |
High Importance
|
In systems where all drives are of type HUSMM80xx0ASS20 it will not be possible to assign a quorum drive
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with drives of type HUSMM80xx0ASS20 |
Trigger |
Attempt to assign drive type as quorum |
Workaround |
Manually assign a different drive type as quorum |
|
7.8.1.8 |
Drives, Quorum |
HU01661 |
All |
High Importance
|
A cache-protection mechanism flag setting can become stuck leading to repeated stops of consistency group synchronisation
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.6 or later using remote copy |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
HyperSwap |
HU01733 |
All |
High Importance
|
Canister information, for the High Density Expansion Enclosure, may be incorrectly reported
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems using the High Density Expansion Enclosure (92F) |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Reliability Availability Serviceability |
HU01797 |
All |
High Importance
|
Hitachi G1500 backend controllers may exhibit higher than expected latency
(show details)
Symptom |
Performance |
Environment |
Systems with Hitachi G1500 backend controllers |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Backend Storage |
HU01813 |
All |
High Importance
|
An issue with Global Mirror stream recovery handling at secondary sites can adversely impact replication performance
(show details)
Symptom |
Performance |
Environment |
Systems using Global Mirror |
Trigger |
Secondary cluster overloaded |
Workaround |
None |
|
7.8.1.8 |
Global Mirror |
HU01824 |
All |
High Importance
|
Switching replication direction for HyperSwap relationships can lead to long I/O timeouts
(show details)
Symptom |
Performance |
Environment |
Systems using HyperSwap |
Trigger |
Switch replication direction of a HyperSwap relationship |
Workaround |
None |
|
7.8.1.8 |
HyperSwap |
HU01839 |
All |
High Importance
|
Where a VMware host is being served volumes, from two different controllers, and an issue, on one controller, causes the related volumes to be taken offline then I/O performance, for the volumes from the other controller, will be adversely affected
(show details)
Symptom |
Performance |
Environment |
Systems running v7.5 or later presenting volumes to VMware hosts, from more than one back-end controller |
Trigger |
Issue on back-end controller takes volumes offline |
Workaround |
None |
|
7.8.1.8 |
Hosts |
HU01842 |
All |
High Importance
|
Bursts of I/O to Read-Intensive Drives can be interpreted as dropped frames against the resident slots, leading to redundant drives being incorrectly failed
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with Read-Intensive Drives |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Drives |
HU01846 |
SVC |
High Importance
|
Silent battery discharge condition will unexpectedly take a SVC node offline putting it into a 572 service state
(show details)
Symptom |
Loss of Redundancy |
Environment |
SVC systems using DH8 & SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Reliability Availability Serviceability |
HU01902 |
V7000, V5000, V3700, V3500 |
High Importance
|
During an upgrade, an issue with VPD migration, can cause a timeout leading to a stalled upgrade
(show details)
Symptom |
Loss of Redundancy |
Environment |
Storwize systems |
Trigger |
Upgrade |
Workaround |
None |
|
7.8.1.8 |
System Update |
HU01276 |
All |
Suggested
|
An issue in the handling of debug data from the FC adapter can cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Reliability Availability Serviceability |
HU01467 |
All |
Suggested
|
Failures in the handling of performance statistics files may lead to missing samples in Spectrum Control and other tools
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
Increase the sampling interval |
|
7.8.1.8 |
System Monitoring |
HU01512 |
All |
Suggested
|
During a DRAID MDisk copy-back operation a miscalculation of the remaining work may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Distributed RAID |
HU01523 |
All |
Suggested
|
An issue with FC adapter initialisation can lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Reliability Availability Serviceability |
HU01556 |
All |
Suggested
|
The handling of memory pool usage by Remote Copy may lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.3 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01564 |
All |
Suggested
|
FlashCopy maps cleaning process is not monitoring the grains correctly which may cause FlashCopy maps to not stop
(show details)
Symptom |
None |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
FlashCopy |
HU01657 |
SVC, V7000, V5000 |
Suggested
|
The 16Gb FC HBA firmware may experience an issue, with the detection of unresponsive links, leading to a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Reliability Availability Serviceability |
HU01715 |
All |
Suggested
|
Issuing a rmvolumecopy command followed by an expandvdisksize command may result in hung I/O leading to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using HyperSwap |
Trigger |
A rmvolumecopy command followed by an expandvdisksize command |
Workaround |
Stop I/O to the volume approx 10 mins prior to expanding it |
|
7.8.1.8 |
HyperSwap |
HU01719 |
All |
Suggested
|
Node warmstart due to a parity error in the HBA driver firmware
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6 and later using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Reliability Availability Serviceability |
HU01751 |
All |
Suggested
|
When RAID attempts to flag a strip as bad, and that strip has already been flagged, a node may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
RAID |
HU01760 |
All |
Suggested
|
FlashCopy map progress appears to be stuck at zero percent
(show details)
Symptom |
None |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
FlashCopy |
HU01786 |
All |
Suggested
|
An issue in the monitoring of SSD write endurance can result in false 1215/2560 errors in the Event Log
(show details)
Symptom |
None |
Environment |
Systems running v7.7.1 or later with SSDs |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Drives |
HU01790 |
All |
Suggested
|
On the Create Volumes page the Accessible I/O Groups selection may not update when the Caching I/O group selection is changed
(show details)
Symptom |
None |
Environment |
Systems with more than one I/O group |
Trigger |
Change "Caching I/O group" selection on "Create Volumes" |
Workaround |
Leave the "Caching I/O group" and "Accessible I/O Groups" selections as "default". Use "modify I/O group" action (right-click "volume" -> "modify I/O group"") to modify volume's iogrp. |
|
7.8.1.8 |
Graphical User Interface |
HU01793 |
All |
Suggested
|
The Maximum final size value in the Expand Volume dialog can display an incorrect value preventing expansion
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
Set volume mirroring memory to minimum. Create a generic mirrored volume of 2TB or less. Open Expand Volume dialog |
Workaround |
Use CLI to expand volume |
|
7.8.1.8 |
Graphical User Interface |
HU02028 |
All |
Suggested
|
An issue, with timer cancellation, in the Remote Copy component may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Remote Copy |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
IT19561 |
All |
Suggested
|
An issue with register clearance in the FC driver code may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Reliability Availability Serviceability |
IT22591 |
All |
Suggested
|
An issue in the HBA adapter firmware may result in node warmstarts
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Reliability Availability Serviceability |
IT24900 |
V7000, V5000 |
Suggested
|
Whilst replacing a control enclosure midplane an issue at boot can prevent VPD being assigned delaying a return to service
(show details)
Symptom |
None |
Environment |
All Storwize systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.8 |
Reliability Availability Serviceability |
IT26836 |
V7000, V5000 |
Suggested
|
Loading drive firmware may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize systems |
Trigger |
Loading drive firmware |
Workaround |
None |
|
7.8.1.8 |
Drives |
HU01802 |
All |
Critical
|
USB encryption key can become inaccessible after upgrade. If the system is later rebooted then any encrypted volumes will be unavailable
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using encryption |
Trigger |
System Upgrade |
Workaround |
None |
|
7.8.1.7 |
Encryption |
HU01785 |
All |
High Importance
|
An issue with memory mapping may lead to multiple node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.7 |
|
HU01792 |
All |
HIPER
|
When a DRAID array has multiple drive failures and the number of failed drives is greater than the number of rebuild areas in the array it is possible that the storage pool will be taken offline during the copyback phase of a rebuild. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
Distributed RAID |
HU01866 |
SVC |
HIPER
|
A faulty PSU sensor, in a node, can fill the sel log causing the service processor (BMC) to disable logging. If a snap is subsequently taken, from the node, a timeout will occur and it will be taken offline. It is possible for this to affect both nodes in an I/O group
(show details)
Symptom |
Loss of Access to Data |
Environment |
SVC systems using SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
System Monitoring |
HU01524 |
All |
Critical
|
When a system loses input power, nodes will shut down until power is restored. If a node was in the process of creating a bad block for an MDisk, at the moment it shuts down, then there is a chance that the system will hit repeated Tier 2 recoveries when it powers back up
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8.1 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
RAID |
HU01767 |
All |
Critical
|
Reads of 4K/8K from an array can under exceptional circumstances return invalid data. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8.0 or earlier |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
RAID, Thin Provisioning |
IT17919 |
All |
Critical
|
A rare timing window issue in the handling of Remote Copy state can result in multi-node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.2 or later using Remote Copy |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01420 |
All |
High Importance
|
An issue in DRAID can cause repeated node warmstarts in the circumstances of a degraded copyback operation to a drive
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
Distributed RAID |
HU01476 |
All |
High Importance
|
A remote copy relationship may suffer a loss of synchronisation when the relationship is renamed
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems using Remote Copy |
Trigger |
Rename remote copy relationship using GUI or CLI |
Workaround |
None |
|
7.8.1.6 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01623 |
All |
High Importance
|
An issue in the handling of inter-node communications can lead to latency for Remote Copy relationships
(show details)
Symptom |
Performance |
Environment |
Systems running v7.1 or later using Remote Copy |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01630 |
All |
High Importance
|
When a system with FlashCopy mappings is upgraded there may be multiple node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.5 or earlier using FlashCopy |
Trigger |
Upgrade to v7.6 or later |
Workaround |
None |
|
7.8.1.6 |
FlashCopy |
HU01697 |
All |
High Importance
|
A timeout issue in RAID member management can lead to multiple node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
RAID |
HU01771 |
SVC, V7000 |
High Importance
|
An issue with the CMOS battery in a node can cause an unexpectedly large log file to be generated by the BMC. At log collection the node may be taken offline
(show details)
Symptom |
Loss of Redundancy |
Environment |
SVC & V7000 systems running v7.8 or later |
Trigger |
Node CMOS battery issue |
Workaround |
None |
|
7.8.1.6 |
System Monitoring |
HU01446 |
All |
Suggested
|
Where host workload overloads the back-end controller and VMware hosts are issuing ATS commands a race condition may be triggered leading to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6.1 or later with VMware hosts using VAAI CAW feature |
Trigger |
None |
Workaround |
Avoid overloading the back-end |
|
7.8.1.6 |
Hosts |
HU01472 |
All |
Suggested
|
A locking issue in Global Mirror can cause a warmstart on the secondary cluster
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Global Mirror |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
Global Mirror |
HU01619 |
All |
Suggested
|
A misreading of the PSU register can lead to failure events being logged incorrectly
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
System Monitoring |
HU01628 |
All |
Suggested
|
In the GUI on the Volumes page whilst using the filter function some volumes entries may not be displayed until the page has completed loading
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
Enter a filter string on the Volumes page |
Workaround |
None |
|
7.8.1.6 |
Graphical User Interface |
HU01664 |
All |
Suggested
|
A timing window issue during an upgrade can cause the node restarting to warmstart stalling the upgrade
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
System Update |
HU01698 |
All |
Suggested
|
A node warmstart may occur when deleting a compressed volume if a host has written to the volume minutes before the volume is deleted
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with compressed volumes |
Trigger |
Write I/O received by compression code for a volume that has been deleted |
Workaround |
Remove volume host mappings and leave an interval of time (90mins) before removing the volume |
|
7.8.1.6 |
Compression |
HU01740 |
All |
Suggested
|
The timeout setting for key server commands may be too brief, when the server is busy, causing those commands to fail
(show details)
Symptom |
None |
Environment |
Systems running v7.8 or later using encryption |
Trigger |
Enter mkkeyserver command |
Workaround |
Retry command |
|
7.8.1.6 |
Encryption |
HU01747 |
All |
Suggested
|
The incorrect detection of a cache issue can lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.6 |
Cache |
HU00247 |
All |
Critical
|
A rare deadlock condition can lead to a RAID5 or RAID6 array rebuild stalling at 99%
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using RAID5/6 |
Trigger |
None |
Workaround |
None |
|
7.8.1.5 |
Distributed RAID, RAID |
HU01620 |
All |
Critical
|
Configuration changes can slow critical processes and, if this coincides with cloud account statistical data being adjusted, a Tier 2 recovery may occur
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8 or later that support TCT |
Trigger |
Configuration changes around midnight |
Workaround |
Avoid making large numbers of configuration changes around midnight |
|
7.8.1.5 |
Transparent Cloud Tiering |
IC57642 |
All |
Critical
|
A complex combination of failure conditions in the fabric connecting nodes can result in lease expiries, possibly cluster-wide
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with more than 2 nodes |
Trigger |
None |
Workaround |
None |
|
7.8.1.5 |
Reliability Availability Serviceability |
IT19192 |
All |
Critical
|
An issue in the handling of GUI certificates may cause warmstarts leading to a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.5 |
Graphical User Interface, Reliability Availability Serviceability |
IT23747 |
All |
High Importance
|
For large drive sizes the DRAID rebuild process can consume significant CPU resource adversely impacting system performance
(show details)
Symptom |
Performance |
Environment |
Systems running v7.6 or later using DRAID |
Trigger |
High write workloads during a DRAID rebuild |
Workaround |
Reduce write workload during rebuild or use a volume/pool level throttle |
|
7.8.1.5 |
Distributed RAID |
HU01655 |
All |
Suggested
|
The algorithm used to calculate an SSDs replacement date can sometimes produce incorrect results leading to a premature End-of-Life error being reported
(show details)
Symptom |
None |
Environment |
Systems using SSDs |
Trigger |
None |
Workaround |
Ignore error if SSD write endurance (in lsdrive CLI output) shows less than 80 |
|
7.8.1.5 |
Drives |
HU01679 |
All |
Suggested
|
An issue in the RAID component can very occasionally cause a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.5 |
RAID |
HU01687 |
All |
Suggested
|
For volumes by host, ports by host and volumes by pool pages in the GUI when the number of items is greater than 50 then the item name will not be displayed
(show details)
Symptom |
None |
Environment |
Systems running v7.7 or v7.8 |
Trigger |
More than 50 hosts/pools |
Workaround |
Use CLI |
|
7.8.1.5 |
Graphical User Interface |
HU01704 |
All |
Suggested
|
In systems using HyperSwap a rare timing window issue can result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using HyperSwap |
Trigger |
None |
Workaround |
None |
|
7.8.1.5 |
HyperSwap |
HU01724 |
All |
Suggested
|
An I/O lock handling issue between nodes can lead to a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.5 |
RAID |
HU01729 |
All |
Suggested
|
Remote copy uses multiple streams to send data between clusters. During a stream disconnect a node, unable to progress, may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Remote Copy |
Trigger |
None |
Workaround |
None |
|
7.8.1.5 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01730 |
SVC |
Suggested
|
When running the DMP for a 1046 error the picture may not indicate the correct position of the failed adapter
(show details)
Symptom |
None |
Environment |
DH8 systems |
Trigger |
Run DMP for 1046 error |
Workaround |
Be aware of which adapter is failed and do not use the picture in the GUI |
|
7.8.1.5 |
GUI Fix Procedure |
HU01731 |
SVC, V7000, V5000 |
Suggested
|
When a node is placed into service mode it is possible for all compression cards within the node to be marked as failed
(show details)
Symptom |
None |
Environment |
Systems running v7.7 or later with compression cards |
Trigger |
Put node in service state |
Workaround |
None |
|
7.8.1.5 |
Compression |
HU01763 |
SVC |
Suggested
|
A single node warmstart may occur on a DH8 config node when inventory email is created. The issue only occurs if this coincides with a very high rate of CLI commands and high I/O workload on the config node
(show details)
Symptom |
Single Node Warmstart |
Environment |
DH8 systems running v7.7.1 or later |
Trigger |
High CLI workload and I/O workload at same time as inventory email |
Workaround |
None |
|
7.8.1.5 |
Command Line Interface, System Monitoring |
IT23140 |
All |
Suggested
|
When viewing the licensed functions GUI page the individual calculations for SCUs, for each tier, may be wrong. However the total is correct
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
Use CLI |
|
7.8.1.5 |
Graphical User Interface |
HU01706 |
All |
HIPER
|
Areas of volumes written with all-zero data may contain non-zero data. For more details refer to this Flash
(show details)
Symptom |
Incorrect data read from volume |
Environment |
Systems running 7.7.1.7 or 7.8.1.3 |
Trigger |
See Flash |
Workaround |
None |
|
7.8.1.4 |
|
HU01646 |
All |
HIPER
|
A new failure mechanism in the 16Gb HBA driver can under certain circumstances lead to a lease expiry of the entire cluster
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with 16Gb HBAs |
Trigger |
Faulty SAN hardware (adapter; SFP; switch) |
Workaround |
None |
|
7.8.1.3 |
Reliability Availability Serviceability |
HU01321 |
All |
Critical
|
Multi-node warmstarts may occur when changing the direction of a remote copy relationship whilst write I/O to the (former) primary volume is still occurring
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using Remote Copy |
Trigger |
Change the direction of a remote copy relationship |
Workaround |
None |
|
7.8.1.3 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01490 |
All |
Critical
|
When attempting to add/remove multiple IQNs to/from a host the tables that record host-wwpn mappings can become inconsistent resulting in repeated node warmstarts across I/O groups
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with iSCSI connected hosts |
Trigger |
- A addhostport command with iqn2 and iqn1 (where iqn1 is already recorded) is entered;
- This command attempts to add iqn2 but determines that iqn1 is a duplicate and the CLI command fails;
- Later whenever a login request from iqn2 is received internal checking detects an inconsistency and warmstarts the node
|
Workaround |
Do not use multiple IQNs in iSCSI add/remove commands |
|
7.8.1.3 |
iSCSI |
HU01549 |
All |
Critical
|
During a system upgrade HyperV-clustered hosts may experience a loss of access to any iSCSI connected volumes
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.5 or earlier with iSCSI connected HyperV clustered hosts |
Trigger |
Upgrade to v7.6 or later |
Workaround |
None |
|
7.8.1.3 |
System Update, iSCSI |
HU01625 |
All |
Critical
|
In systems with a consistency group of HyperSwap or Metro Mirror relationships if an upgrade attempts to commit whilst a relationship is out of synch then there may be multiple warmstarts and a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7 or earlier using HyperSwap or Metro Mirror |
Trigger |
Upgrading to v7.8 or later |
Workaround |
Stop all replication prior to upgrading |
|
7.8.1.3 |
System Update, HyperSwap, Metro Mirror |
IT23034 |
All |
Critical
|
With HyperSwap volumes and mirrored copies, at a single site, using rmvolumecopy to remove a copy, from an auxiliary volume, may result in a cluster-wide warmstart necessitating a Tier 2 recovery
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems with HyperSwap volumes and mirrored copies at a single site |
Trigger |
Using rmvolumecopy to remove a copy from a HyperSwap auxiliary volume |
Workaround |
None |
|
7.8.1.3 |
HyperSwap |
HU01481 |
All |
High Importance
|
A failed I/O can trigger HyperSwap to unexpectedly change the direction of the relationship leading to node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.5 or later using HyperSwap |
Trigger |
None |
Workaround |
None |
|
7.8.1.3 |
HyperSwap |
HU01525 |
All |
High Importance
|
During an upgrade a resource locking issue in the compression component can cause a node to warmstart multiple times and become unavailable
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.6 or earlier using compressed volumes |
Trigger |
Upgrade to v7.7 or later |
Workaround |
None |
|
7.8.1.3 |
Compression, System Update |
HU01569 |
SVC |
High Importance
|
When compression utilisation is high the config node may exhibit longer I/O response times than non-config nodes
(show details)
Symptom |
Performance |
Environment |
SVC systems using compression |
Trigger |
High compression workloads |
Workaround |
None |
|
7.8.1.3 |
Compression |
HU01584 |
All |
High Importance
|
An issue in array indexing can cause a RAID array to go offline repeatedly
(show details)
Symptom |
Offline Volumes |
Environment |
System running v7.6 or later |
Trigger |
None |
Workaround |
Avoid doing member exchanges |
|
7.8.1.3 |
RAID |
HU01614 |
All |
High Importance
|
After a node is upgraded hosts defined as TPGS may have paths set to inactive
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.6 or earlier with host type TPGS |
Trigger |
Upgrade to v7.7 or later |
Workaround |
None |
|
7.8.1.3 |
Hosts |
HU01632 |
All |
High Importance
|
A congested fabric causes the Fibre Channel adapter firmware to abort I/O resulting in node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All systems |
Trigger |
Congested fabric |
Workaround |
Avoid fabric congestion |
|
7.8.1.3 |
Reliability Availability Serviceability |
HU01636 |
V5000, V3700, V3500 |
High Importance
|
A connectivity issue with certain host SAS HBAs can prevent hosts from establishing stable communication with the storage controller
(show details)
Symptom |
Performance |
Environment |
Systems presenting storage to hosts with N2225 adapters |
Trigger |
Host with N2225 adapters running Windows Server 2012R2 |
Workaround |
None |
|
7.8.1.3 |
Hosts |
HU01638 |
All |
High Importance
|
When upgrading to v7.6 or later, if there is another cluster in the same zone which is at v5.1 or earlier then nodes will warmstart and the upgrade will fail
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.5 or earlier |
Trigger |
Upgrade to v7.6 or later with a cluster in the same zone running v6.1 or earlier |
Workaround |
Unzone any cluster running v5.1 or earlier from the cluster being upgraded |
|
7.8.1.3 |
System Update |
HU01645 |
SVC |
High Importance
|
After upgrading to v7.8 a reboot of a node will initiate a continual boot cycle
(show details)
Symptom |
Loss of Redundancy |
Environment |
DH8 systems running v7.7.0 or earlier with 32GB of RAM |
Trigger |
Upgrade to v7.8; Reboot node |
Workaround |
None |
|
7.8.1.3 |
System Update |
HU01385 |
All |
Suggested
|
A warmstart may occur if a rmvolumecopy or rmrcrelationship command are issued on a volume while I/O is being forwarded to the associated copy
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6 or v7.7 that are using Hyperswap |
Trigger |
Issue an rmvolumecopy or rmrcrelationship command whilst hosts are still actively using the Hyperswap volume |
Workaround |
Do not remove a Hyperswap volume or relationship whilst hosts are still mapped to it |
|
7.8.1.3 |
HyperSwap |
HU01457 |
V7000 |
Suggested
|
In a hybrid V7000 cluster where one I/O group supports 10k volumes and another does not some operations on volumes may incorrectly be denied in the GUI
(show details)
Symptom |
None |
Environment |
Systems running v7.7.1 or later |
Trigger |
None |
Workaround |
Performed required actions in the CLI |
|
7.8.1.3 |
Graphical User Interface |
HU01535 |
All |
Suggested
|
An issue with Fibre Channel driver handling of command processing can result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6 or later using Fibre Channel connectivity |
Trigger |
None |
Workaround |
None |
|
7.8.1.3 |
|
HU01563 |
V7000 |
Suggested
|
Where an IBM SONAS host id is used it can under rare circumstances cause a warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Unified configurations |
Trigger |
None |
Workaround |
None |
|
7.8.1.3 |
|
HU01582 |
All |
Suggested
|
A compression issue in IP replication can result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7 or later using IP Replication |
Trigger |
None |
Workaround |
None |
|
7.8.1.3 |
IP Replication |
HU01624 |
All |
Suggested
|
GUI response can become very slow in systems with a large number of compressed and uncompressed volume
(show details)
Symptom |
None |
Environment |
Systems that have a large number of volumes (1000+) including some that are compressed |
Trigger |
None |
Workaround |
None |
|
7.8.1.3 |
Graphical User Interface |
HU01631 |
All |
Suggested
|
A memory leak in EasyTier when pools are in Balanced mode can lead to node warmstarts
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later using EasyTier |
Trigger |
Pools in Balanced mode |
Workaround |
Create hybrid pools with multiple tiers of storage and set Easy tier mode to Active |
|
7.8.1.3 |
EasyTier |
HU01654 |
SVC, V7000, V5000 |
Suggested
|
There may be a node warmstart when a switch of direction, in a HyperSwap relationship, fails to complete properly
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later using HyperSwap |
Trigger |
None |
Workaround |
None |
|
7.8.1.3 |
HyperSwap |
HU01239 & HU01255 & HU01586 |
All |
HIPER
|
The presence of a faulty SAN component can delay lease messages between nodes leading to a cluster-wide lease expiry and consequential loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with 16Gb HBAs |
Trigger |
Faulty SAN hardware (adapter, SFP, switch) |
Workaround |
None |
|
7.8.1.2 |
Reliability Availability Serviceability |
HU01626 |
All |
High Importance
|
Node downgrade from v7.8.x to v7.7.1 or earlier (e.g. during an aborted upgrade) may prevent the node from rejoining the cluster. Systems that have already completed upgrade to v7.8.x are not affected by this issue
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.8 or later |
Trigger |
Downgrade to v7.7.1 or earlier |
Workaround |
None |
|
7.8.1.2 |
System Update |
HU01505 |
All |
HIPER
|
A non-redundant drive experiencing many errors can be taken offline obstructing rebuild activity
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Drive with many errors |
Workaround |
Bring drive online |
|
7.8.1.1 |
Backend Storage, RAID |
HU01570 |
V7000, V5000, V3700, V3500 |
HIPER
|
Reseating a drive in an array may cause the MDisk to go offline
(show details)
Symptom |
Offline Volumes |
Environment |
Storwize Gen 1 systems running v7.8 or later |
Trigger |
Removing a drive at the point of failure |
Workaround |
Only remove a drive when offline/failed |
|
7.8.1.1 |
Drives, RAID |
IT20627 |
All |
Critical
|
When Read-Intensive drives are used as quorum disks a drive outage can occur. Under some circumstances this can lead to a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Using Read-Intensive drives as quorum disks |
Workaround |
None |
|
7.8.1.1 |
Quorum |
HU01477 |
V7000, V5000, V3700, V3500 |
High Importance
|
Due to the way enclosure data is read it is possible for a firmware mismatch between nodes to occur during an upgrade
(show details)
Symptom |
Loss of Redundancy |
Environment |
Storwize systems running v7.8 or later |
Trigger |
Upgrade from v7.8.0 to v7.8.1 |
Workaround |
Check for dependent volumes. If there are none then manually load enclosure firmware from service mode |
|
7.8.1.1 |
System Update |
HU01503 |
All |
High Importance
|
When the 3PAR host type is set to legacy the round robin algorithm, used to select the MDisk port for I/O submission to 3PAR controllers, does not work correctly and I/O may be submitted to fewer controller ports, adversely affecting performance
(show details)
Symptom |
Performance |
Environment |
Systems running v7.5.0.10 v7.6.1.6 or v7.7.1.5 that are virtualising 3PAR storage subsystems |
Trigger |
Host persona set to 6 (legacy) on 3PAR controller |
Workaround |
Change 3PAR host persona to 2 (ALUA) instead of 6 (legacy) |
|
7.8.1.1 |
Backend Storage |
HU01609 & IT15343 |
All |
High Importance
|
When the system is busy, the compression component may be paged out of memory resulting in latency that can lead to warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using compression |
Trigger |
None |
Workaround |
Reduce compression workload |
|
7.8.1.1 |
Compression |
IT19726 |
SVC |
High Importance
|
Warmstarts may occur when the attached SAN fabric is congested and HBA transmit paths become stalled preventing the HBA firmware from generating the completion for a FC command
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
SVC systems |
Trigger |
SAN congestion |
Workaround |
None |
|
7.8.1.1 |
Hosts |
HU00763 & HU01237 |
V7000, V5000, V3700, V3500 |
Suggested
|
A node warmstart may occur when a quorum disk is accessed at the same time as the login to that disk is closed
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.1 |
Quorum |
HU01332 |
All |
Suggested
|
Performance monitor and Spectrum Control show zero CPU utilisation for compression
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.1 |
System Monitoring |
HU01353 |
All |
Suggested
|
CLI allows the input of carriage return characters into certain fields, after cluster creation, resulting in invalid cluster VPD and failed node adds
(show details)
Symptom |
Configuration |
Environment |
All systems |
Trigger |
After cluster creation use the CLI to enter a carriage return in a command that allows free text in an argument |
Workaround |
Do not insert a carriage return character into text being entered via CLI |
|
7.8.1.1 |
Command Line Interface |
HU01391 & HU01581 |
V7000, V5000, V3700, V3500 |
Suggested
|
Storwize systems may experience a warmstart due to an uncorrectable error in the SAS firmware
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.1 |
Drives |
HU01430 |
V7000, V5000, V3700, V3500 |
Suggested
|
Memory resource shortages in systems with 8GB of RAM can lead to node warmstarts
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize Gen 1 systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.1 |
|
HU01469 |
V3700, V3500 |
Suggested
|
Resource exhaustion in the iSCSI component can result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize V3k systems using iSCSI |
Trigger |
None |
Workaround |
None |
|
7.8.1.1 |
iSCSI |
HU01471 |
V5000 |
Suggested
|
Power system down using the GUI on V5000 causes the fans to run high while the system is offline but power is still applied to the enclosure
(show details)
Symptom |
None |
Environment |
Storwize V5k systems running v7.6 or later |
Trigger |
svctask stopsystem -gui |
Workaround |
After shutdown remove power from enclosure |
|
7.8.1.1 |
|
HU01484 |
All |
Suggested
|
During a RAID array rebuild there may be node warmstarts
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.1 |
RAID |
HU01496 |
SVC |
Suggested
|
SVC node type SV1 reports wrong FRU part number for compression accelerator
(show details)
Symptom |
None |
Environment |
SVC SV1 systems |
Trigger |
Run lsnodevpd |
Workaround |
If a replacement is required order FRU part number 01EJ817 |
|
7.8.1.1 |
Command Line Interface |
HU01520 |
V3700, V3500 |
Suggested
|
Where the system is being used as secondary site for Remote Copy during an upgrade to v7.8.1 the node may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize V3500 & V3700 systems running v7.1 or later being used as a secondary site |
Trigger |
Upgrade to v7.8.1 |
Workaround |
Stop all replication prior to an upgrade to v7.8.1 |
|
7.8.1.1 |
System Update, Global Mirror, Metro Mirror |
HU01531 |
All |
Suggested
|
Spectrum Control is unable to receive notifications from SVC/Storwize. Spectrum Control may experience an out-of-memory condition
(show details)
Symptom |
None |
Environment |
Systems running v7.8.1.0 that are being managed by IBM Spectrum Control |
Trigger |
None |
Workaround |
None |
|
7.8.1.1 |
System Monitoring |
HU01566 |
SVC |
Suggested
|
After upgrading, numerous 1370 errors are seen in the Event Log
(show details)
Symptom |
None |
Environment |
SVC CG8 systems containing old SSD drives (type Z16IZD2B-146) |
Trigger |
System Upgrade |
Workaround |
Ignore 1370 errors |
|
7.8.1.1 |
System Update |
IT19973 |
All |
Suggested
|
Call home emails may not be sent due to a failure to retry
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.1 |
|
HU01474 |
SVC, V7000 |
HIPER
|
Host writes to a read-only secondary volume trigger I/O timeout warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
SVC SV1 & V7000 Gen 2+ systems running v7.7 or later using Remote Copy |
Trigger |
Host write I/O to a read-only secondary volume |
Workaround |
Do not attempt to issue host writes to a secondary volume while remote copy is active |
|
7.8.1.0 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01479 |
All |
HIPER
|
The handling of drive reseats can sometimes allow I/O to occur before the drive has been correctly failed resulting in offline MDisks
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.6 or later using DRAID |
Trigger |
Reseat a failed drive |
Workaround |
Rather than reseating the drive use the CLI or GUI to fail then unfail it |
|
7.8.1.0 |
Distributed RAID |
HU01483 |
All |
HIPER
|
mkdistributedarray command may get stuck in the prepare state. Any interaction with the volumes in that array will result in multiple warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1 or later |
Trigger |
mkdistributedarray command gets stuck in the prepare state; run rmarray or rmvdisk commands or generate I/O to volumes created on that array |
Workaround |
None |
|
7.8.1.0 |
Distributed RAID |
HU01675 |
V7000 |
HIPER
|
Memory allocation issues may cause GUI and I/O performance issues
(show details)
Symptom |
Loss of Access to Data |
Environment |
V7000 Gen2/2+ systems running v7.7.1.7 or later using compression |
Trigger |
Upgrade to v7.7.1.7 or later |
Workaround |
None |
|
7.8.1.0 |
Compression, Graphical User Interface |
HU01220 |
All |
Critical
|
Changing the type of a RC consistency group when a volume in a subordinate relationship is offline will cause a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.5 or later using GMCV |
Trigger |
GMCV relationship with offline volume; Change overarching consistency group type from GMCV to GM |
Workaround |
Identify relationships with offline change volumes and remove them from the consistency group before changing the group's type. Change the type of the relationship separately and then add it back into the group. |
|
7.8.1.0 |
Global Mirror with Change Volumes, Global Mirror |
HU01252 |
SVC |
Critical
|
Where a SVC is presenting storage from an 8-node V7000, an upgrade to that V7000 can pause I/O long enough for the SVC to take related MDisks offline
(show details)
Symptom |
Offline Volumes |
Environment |
SVC systems with large and complex V7000 configurations |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
|
HU01416 |
All |
Critical
|
ISL configuration activity may cause a cluster-wide lease expiry
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
Reliability Availability Serviceability |
HU00747 |
V7000, V5000, V3700, V3500 |
High Importance
|
Node warmstarts can occur when drives become degraded
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Storwize systems running v7.3 or later |
Trigger |
Degraded drives |
Workaround |
Avoid degraded drives |
|
7.8.1.0 |
Backend Storage |
HU01309 |
SVC |
High Importance
|
For FC logins, on a node that is online for more than 200 days, if a fabric event makes a login inactive then the node may be unable to re-establish the login
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with 8GB FC cards |
Trigger |
FC login established; node up for >200 days; fabric event causes login to go inactive |
Workaround |
None |
|
7.8.1.0 |
Backend Storage |
HU01371 |
SVC, V7000, V5000 |
High Importance
|
A remote copy command related to HyperSwap may hang resulting in a warmstart of the config node
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.5 or later using HyperSwap |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
HyperSwap |
HU01388 |
SVC, V7000, V5000 |
High Importance
|
Where a HyperSwap volume is the source of a FlashCopy mapping and the HyperSwap relationship is out of sync when the HyperSwap volume comes back online a switch of direction will occur and FlashCopy operation may delay I/O leading to node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using FlashCopy with HyperSwap |
Trigger |
An offline HyperSwap volume that is the source of a FlashCopy mapping is brought back online |
Workaround |
Stop dependent FlashCopy mappings or let them complete before bringing an offline HyperSwap volume back online |
|
7.8.1.0 |
HyperSwap, FlashCopy |
HU01394 |
All |
High Importance
|
Node warmstarts may occur on systems which are performing Global Mirror replication, due to a low-probability timing window
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.2 or later using GM |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
Global Mirror |
HU01395 |
All |
High Importance
|
Malformed URLs sent by security scanners whilst correctly discarded can cause considerable exception logging on config nodes leading to performance degradation that can adversely affect remote copy
(show details)
Symptom |
Performance |
Environment |
Systems using GM |
Trigger |
Security scanning config node |
Workaround |
Avoid security scanning Storwize clusters |
|
7.8.1.0 |
Global Mirror |
HU01413 |
All |
High Importance
|
Node warmstarts when establishing an FC partnership between a system on v7.7.1 or later with another system which in turn has a partnership to another system running v6.4.1 or earlier
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.7.1 or later with partnerships to systems which have partnerships with systems running v6.4.1 or earlier |
Trigger |
Create FC partnership |
Workaround |
Upgrade systems on pre-7.2 code levels |
|
7.8.1.0 |
Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01428 |
V7000, V5000, V3700, V3500 |
High Importance
|
Scheduling issue adversely affects performance resulting in node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Storwize systems running v7.7.1 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
Reliability Availability Serviceability |
HU01480 |
All |
High Importance
|
Under some circumstances the config node does not fail over properly when using IPv6 adversely affecting management access via GUI and CLI
(show details)
Symptom |
Configuration |
Environment |
Systems using IPv6 cluster addresses |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
Command Line Interface, Graphical User Interface |
IT19019 |
V5000 |
High Importance
|
V5000 control enclosure midplane FRU replacement may fail leading to both nodes reporting a 506 error
(show details)
Symptom |
None |
Environment |
V5000 systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
Reliability Availability Serviceability |
HU01057 |
SVC |
Suggested
|
Slow GUI performance for some pages as the lsnodebootdrive command generates unexpected output
(show details)
Symptom |
Performance |
Environment |
SVC systems with CG8 model nodes or older that are running v7.3 or later |
Trigger |
Select Monitor -> Systems page |
Workaround |
Remove and re-add non-DH8 nodes |
|
7.8.1.0 |
Graphical User Interface |
HU01404 |
All |
Suggested
|
A node warmstart may occur when a new volume is created using fast format and foreground I/O is submitted to the volume
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
Volume is actively fast formatting when write I/O arrives |
Workaround |
Do not use fast format, or if fast format is used do not submit write I/O to the volume until the format process is complete |
|
7.8.1.0 |
|
HU01445 |
SVC, V7000 |
Suggested
|
Systems with heavily used RAID-1 or RAID-10 arrays may experience a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
SVC & V7000 systems running v7.7 or later |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
|
HU01463 |
All |
Suggested
|
SSH Forwarding is enabled on the SSH server
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
|
HU01466 |
SVC, V7000, V5000 |
Suggested
|
Stretched cluster and HyperSwap I/O routing does not work properly due to incorrect ALUA data
(show details)
Symptom |
None |
Environment |
Systems running v7.8 or later using HyperSwap |
Trigger |
Move a host to an alternate I/O group |
Workaround |
Use chhost -site to set the correct site |
|
7.8.1.0 |
Hosts, HyperSwap |
HU01470 |
All |
Suggested
|
T3 might fail during svcconfig recover -execute while running chemail if the email_machine_address contains a comma
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
email_machine_address contains a comma with a preceding whitespace |
Workaround |
Modify the config backup or svc.config.recover.sh script by removing trailing whitespace |
|
7.8.1.0 |
Reliability Availability Serviceability |
HU01473 |
All |
Suggested
|
EasyTier migrates an excessive number of cold extents to an overloaded nearline array
(show details)
Symptom |
Performance |
Environment |
Systems running v7.6.1 or later using EasyTier |
Trigger |
None |
Workaround |
None |
|
7.8.1.0 |
EasyTier |
HU01487 |
All |
Suggested
|
Small increase in read response time for source volumes with additional FlashCopy maps
(show details)
Symptom |
Performance |
Environment |
Systems running v7.5 or later using Global Mirror with Change Volumes |
Trigger |
Add FlashCopy mappings to GMCV source volumes |
Workaround |
Avoid using incremental FlashCopy |
|
7.8.1.0 |
FlashCopy, Global Mirror with Change Volumes |
HU01497 |
All |
Suggested
|
A drive can still be offline even though the error is showing as corrected in the Event Log
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later using DRAID |
Trigger |
None |
Workaround |
Reseat/replace drive |
|
7.8.1.0 |
Distributed RAID |
HU01498 |
All |
Suggested
|
GUI may be exposed to CVE-2017-5638 (see Section 3.1)
|
7.8.1.0 |
Security |
IT19232 |
V7000, V5000, V3700, V3500 |
Suggested
|
Storwize systems can report unexpected drive location errors as a result of a RAID issue
(show details)
Symptom |
None |
Environment |
Storwize systems running v7.3 or later |
Trigger |
None |
Workaround |
Replace the drive |
|
7.8.1.0 |
|
HU01225 & HU01330 & HU01412 |
All |
Critical
|
Node warmstarts due to inconsistencies arising from the way cache interacts with compression
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.3 or later with compressed volumes |
Trigger |
None |
Workaround |
None |
|
7.8.0.2 |
Cache, Compression |
HU01410 |
SVC |
Critical
|
An issue in the handling of FlashCopy map preparation can cause both nodes in an I/O group to be put into service state
(show details)
Symptom |
Loss of Access to Data |
Environment |
SVC systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.8.0.2 |
FlashCopy |
HU01442 |
All |
Critical
|
Upgrading to v7.7.1.5 or v7.8.0.1 with encryption enable will result in multiple Tier 2 recoveries and a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.5 or earlier |
Trigger |
Upgrade to v7.8.0.1 |
Workaround |
None |
|
7.8.0.2 |
Encryption, System Update |
HU00762 |
All |
High Importance
|
Due to an issue in the cache component, nodes within an I/O group are not able to form a caching-pair and are serving I/O through a single node
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.3 or later |
Trigger |
None |
Workaround |
None |
|
7.8.0.2 |
Cache |
HU01409 |
All |
High Importance
|
Cisco Nexus 3000 switches at v5.0(3) have a defect which prevents a config node IP address changing in the event of a fail over
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems connected to Cisco Nexus 3000 switches |
Trigger |
Config node fail over |
Workaround |
None |
|
7.8.0.2 |
Reliability Availability Serviceability |
HU01426 |
All |
High Importance
|
Systems running v7.6.1 or earlier, with compressed volumes, that upgrade to v7.8.0 or later will fail when the first node warmstarts and enters a service state
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.6.1 or earlier using compressed volumes |
Trigger |
Upgrade from v7.6.1 or earlier to 7.8.0.0 or later with compressed volumes |
Workaround |
None |
|
7.8.0.2 |
System Update |
HU01432 |
All |
Suggested
|
Node warmstart due to an accounting issue within the cache component
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.3 or later |
Trigger |
None |
Workaround |
None |
|
7.8.0.2 |
Cache |
HU01459 |
V7000, V5000 |
Suggested
|
The event log indicates incorrect enclosure type
(show details)
Symptom |
None |
Environment |
Storwize V5000 and V7000 Gen 2 systems |
Trigger |
None |
Workaround |
Replace the enclosure with a recently manufactured FRU part |
|
7.8.0.2 |
System Monitoring |
IT18752 |
All |
Suggested
|
When the config node processes an lsdependentvdisks command, issued via the GUI, that has a large number of objects in its parameters, it may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7 or later |
Trigger |
Issue lsdependentvdisks command, via the GUI, for a large number of objects |
Workaround |
Issue those commands, via the CLI |
|
7.8.0.2 |
Graphical User Interface |
HU01382 |
All |
HIPER
|
Mishandling of extent migration following a rmarray command can lead to multiple simultaneous node warmstarts with a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1 or later using DRAID |
Trigger |
Executing a rmarray command |
Workaround |
Avoid using rmarray command |
|
7.8.0.1 |
Distributed RAID |
HU01415 |
V3700 |
Critical
|
When a V3700 with 1GE adapters is upgraded to v7.8.0.0 iSCSI hosts will lose access to volumes
(show details)
Symptom |
Loss of Access to Data |
Environment |
V3700 systems running v7.8.0.0 with 1GE adapters |
Trigger |
Upgrade to v7.8.0.0 |
Workaround |
None |
|
7.8.0.1 |
iSCSI, Hosts |
HU01193 |
All |
HIPER
|
A drive failure whilst an array rebuild is in progress can lead to both nodes in an I/O group warmstarting
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
Drive failure |
Workaround |
None |
|
7.8.0.0 |
Distributed RAID |
HU01379 |
All |
HIPER
|
Resource leak in the handling of Read Intensive drives leads to offline volumes
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v7.7.1 or later using read intensive drives |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Drives |
HU00906 |
All |
Critical
|
When a compressed volume mirror copy is taken offline, write response times to the primary copy may reach prohibitively high levels leading to a loss of access to that volume
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.3 or later that are using volume mirroring with compression |
Trigger |
None |
Workaround |
Only operate mirrored copies on non-compressed volumes |
|
7.8.0.0 |
Compression, Volume Mirroring |
HU01021 & HU01157 |
All |
Critical
|
A fault in a backend controller can cause excessive path state changes leading to node warmstarts and offline volumes
(show details)
Symptom |
Offline Volumes |
Environment |
Systems with backend storage |
Trigger |
Faulty backend controller adapters |
Workaround |
None |
|
7.8.0.0 |
Backend Storage |
HU01267 |
All |
Critical
|
An unusual interaction between Remote Copy and FlashCopy can lead to both nodes in an I/O group warmstarting
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using GMCV |
Trigger |
There may be a number of different scenarios that trigger this issue. Here is one: With GMCV running; Stop GMCV with -access (i.e. svctask stoprcconsistgrp -access); Secondary site becomes primary site and vice versa; Start a FlashCopy on the now primary site; Stop FlashCopy restore before it completes (with -force); Current primary becomes exposed to the issue; Reinstate GMCV (primary becomes secondary again); Stop GMCV with -access again; Both nodes warmstart. |
Workaround |
None |
|
7.8.0.0 |
Global Mirror with Change Volumes |
HU01320 |
All |
Critical
|
A rare timing condition can cause hung I/O leading to warmstarts on both nodes in an I/O group. Probability can be increased in the presence of failing drives.
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
Replace faulty drives promptly |
|
7.8.0.0 |
Hosts |
HU01340 |
All |
Critical
|
A port translation issue between v7.5 or earlier and v7.7.0 or later requires a Tier 2 recovery to complete an upgrade
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.5 or earlier |
Trigger |
Upgrade to v7.7.x |
Workaround |
None |
|
7.8.0.0 |
System Update |
HU01392 |
All |
Critical
|
Under certain rare conditions FC mappings not in a consistency group can be added to a special internal consistency group resulting in a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1 or later |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
FlashCopy |
HU01455 |
All |
Critical
|
VMWare hosts with ATS enabled can see LUN disconnects to volumes when GMCV is used
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7 or later presenting volumes to VMware host using Global Mirror with Change Volumes |
Trigger |
Enable ATS |
Workaround |
Disable ATS when using Global Mirror with Change Volumes |
|
7.8.0.0 |
Global Mirror with Change Volumes |
HU01519 |
V7000 |
Critical
|
One PSU may silently fail leading to the possibility of a dual node reboot
(show details)
Symptom |
Loss of Access to Data |
Environment |
Storwize V7000 Gen 1 systems |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Reliability Availability Serviceability |
HU01635 |
All |
Critical
|
A slow memory leak in the host layer can lead to an out-of-memory condition resulting in offline volumes or performance degradation
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Hosts, Performance |
HU01783 |
All |
Critical
|
Replacing a failed drive in a DRAID array, with a smaller drive, may result in multiple Tier 2 recoveries putting all nodes in service state with error 564 and/or 550
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
Replacing a drive with one of less capacity |
Workaround |
Ensure replacement drives are the same capacity |
|
7.8.0.0 |
Distributed RAID |
HU01831 |
All |
Critical
|
Cluster-wide warmstarts may occur when the SAN delivers a FDISC frame with an invalid WWPN
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.6 or later |
Trigger |
FDISC frame with an invalid WWPN arrives from SAN |
Workaround |
None |
|
7.8.0.0 |
Reliability Availability Serviceability |
HU01177 |
All |
High Importance
|
A small timing window issue exists where a node warmstart or power failure can lead to repeated warmstarts of that node until a node rescue is performed
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Reliability Availability Serviceability |
HU01223 |
All |
High Importance
|
The handling of a rebooted nodes return to the cluster can occasionally become delayed resulting in a stoppage of inter-cluster relationships
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.3 or later using MetroMirror |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Metro Mirror |
HU01254 |
SVC |
High Importance
|
A fluctuation of input AC power can cause a 584 error on a node
(show details)
Symptom |
Loss of Redundancy |
Environment |
SVC CF8/CG8 systems |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Reliability Availability Serviceability |
HU01268 |
V7000, V5000, V3700, V3500 |
High Importance
|
Upgrade to 7.7.x fails on Storwize systems in the replication layer where a T3 recovery was performed in the past
(show details)
Symptom |
None |
Environment |
Storwize systems in the replication layer that have experienced a T3 recovery in the past |
Trigger |
Upgrade to v7.7.x |
Workaround |
None |
|
7.8.0.0 |
System Update |
HU01347 |
All |
High Importance
|
During an upgrade to v7.7.1 a deadlock in node communications can occur leading to a timeout and node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.7.0 or earlier using SE volumes |
Trigger |
Upgrade to v7.7.1 |
Workaround |
None |
|
7.8.0.0 |
Thin Provisioning |
HU01381 |
All |
High Importance
|
A rare timing issue in FlashCopy may lead to a node warmstarting repeatedly and then entering a service state
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.2 or later using FlashCopy |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
FlashCopy |
HU01402 |
V7000 |
High Importance
|
Nodes can power down unexpectedly as they are unable to determine from their partner whether power is available
(show details)
Symptom |
Loss of Redundancy |
Environment |
V7000 Gen 1 systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Reliability Availability Serviceability |
HU01488 |
V7000, V5000 |
High Importance
|
SAS transport errors on an enclosure slot have the potential to affect an adjacent slot leading to double drive failures
(show details)
Symptom |
Loss of Redundancy |
Environment |
Storwize V5000 and V7000 systems running v7.7 or later |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Drives |
IT14917 |
All |
High Importance
|
Node warmstarts due to a timing window in the cache component. For more details refer to this Flash
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v7.4 or later |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Cache |
IT16012 |
SVC |
High Importance
|
Internal node boot drive RAID scrub process at 1am every Sunday can impact system performance
(show details)
Symptom |
Performance |
Environment |
Systems running v7.3 or later |
Trigger |
Internal node boot drive RAID scrub process |
Workaround |
Try to avoid performing high I/O workloads (including copy services) at 1am on Sundays |
|
7.8.0.0 |
Performance |
IT17564 |
All |
High Importance
|
All nodes in an I/O group may warmstart when a DRAID array experiences drive failures
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with DRAID arrays |
Trigger |
Failing drives |
Workaround |
None |
|
7.8.0.0 |
Distributed RAID |
HU00831 |
All |
Suggested
|
Single node warmstart due to hung I/O caused by cache deadlock
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.2 or later |
Trigger |
Hosts sending many large block IOs that use many credits per I/O |
Workaround |
None |
|
7.8.0.0 |
Cache |
HU01098 |
All |
Suggested
|
Some older backend controller code levels do not support C2 commands resulting in 1370 entries in the Event Log for every detectmdisk
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later attached to backend controllers running older code levels |
Trigger |
Issue CLI command detectmdisk |
Workaround |
None |
|
7.8.0.0 |
Backend Storage |
HU01213 |
All |
Suggested
|
The LDAP password is visible in the auditlog
(show details)
Symptom |
None |
Environment |
Systems running v7.7 or later |
Trigger |
Setup LDAP authentication |
Workaround |
None |
|
7.8.0.0 |
LDAP |
HU01228 |
All |
Suggested
|
Automatic T3 recovery may fail due to the handling of quorum registration generating duplicate entries
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Reliability Availability Serviceability |
HU01229 |
V7000, V5000, V3700, V3500 |
Suggested
|
The DMP for a 3105 event does not identify the correct problem canister
(show details)
Symptom |
None |
Environment |
Storwize systems |
Trigger |
None |
Workaround |
Problem canister identified in Event log |
|
7.8.0.0 |
GUI Fix Procedure |
HU01230 |
All |
Suggested
|
A host aborting an outstanding logout command can lead to a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
|
HU01247 |
All |
Suggested
|
When a FlashCopy consistency group is stopped more than once in rapid succession a node warmstart may result
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using FlashCopy |
Trigger |
Stop the same FlashCopy consistency group twice in rapid succession using -force on second attempt |
Workaround |
Avoid stopping the same flash copy consistency group more than once |
|
7.8.0.0 |
FlashCopy |
HU01264 |
All |
Suggested
|
Node warmstart due to an issue in the compression optimisation process
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems that are using compressed volumes |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Compression |
HU01269 |
All |
Suggested
|
A rare timing conflict between two process may lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
|
HU01304 |
All |
Suggested
|
SSH authentication fails if multiple SSH keys are configured on the client
(show details)
Symptom |
None |
Environment |
Systems running v7.7 or later |
Trigger |
On client create 2 x SSH key pairs (neither of which are configured on the SVC/Storwize) then connect to SVC/Storwize system |
Workaround |
Remove other keys from the client or explicitly use password authentication on the client: ssh -o PreferredAuthentications=password <user>@<host> |
|
7.8.0.0 |
|
HU01323 |
All |
Suggested
|
Systems using Volume Mirroring that upgrade to v7.7.1.x and have a storage pool go offline may experience a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7.1 or later using volume mirroring |
Trigger |
1. Mirrored volumes are created at a pre-7.7.1 code level; 2. The system is upgraded to 7.7.1.1 or later; 3. A storage pool goes offline before the nodes next restart. |
Workaround |
None |
|
7.8.0.0 |
Volume Mirroring |
HU01370 |
All |
Suggested
|
lsfabric command may not list all logins when it is used with parameters
(show details)
Symptom |
None |
Environment |
Systems running v7.7 or later |
Trigger |
Enter lsfabric with parameters |
Workaround |
Enter lsfabric without parameters and use client tools such as grep for filtering |
|
7.8.0.0 |
Command Line Interface |
HU01374 |
All |
Suggested
|
Where an issue with Global Mirror causes excessive I/O delay, a timeout may not function resulting in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.7.0 or later using Global Mirror |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
Global Mirror |
HU01399 |
All |
Suggested
|
For certain config nodes the CLI Help commands may not work
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
Use Knowledge Center |
|
7.8.0.0 |
Command Line Interface |
IT17302 |
V5000, V3700, V3500 |
Suggested
|
Unexpected 45034 1042 entries in the Event Log
(show details)
Symptom |
None |
Environment |
Systems running v7.7 or later |
Trigger |
None |
Workaround |
None |
|
7.8.0.0 |
System Monitoring |
IT18086 |
All |
Suggested
|
When a volume is moved between I/O groups a node may warmstart
(show details)
Symptom |
None |
Environment |
Systems running v7.7.1 or later |
Trigger |
No warmstarts since upgrade. Move a volume between I/O groups |
Workaround |
None |
|
7.8.0.0 |
|