Release Note for systems built with IBM Spectrum Virtualize


This is the release note for the 7.8 release and details the issues resolved in all Program Temporary Fixes (PTFs) between 7.8.0.0 and 7.8.1.16. This document will be updated with additional information whenever a PTF is released.

This document was last updated on 1 November 2022.

  1. New Features
  2. Known Issues and Restrictions
  3. Issues Resolved
    1. Security Issues Resolved
    2. APARs Resolved
  4. Useful Links
Note. Detailed build version numbers are included in the Update Matrices in the Useful Links section

1. New Features

The following new features have been introduced in the 7.8.0 release: The following new features have been introduced in the 7.8.1 release: The following feature has been introduced in the 7.8.1.9 release:

2. Known Issues and Restrictions

Note: For clarity, the term "node" will be used to refer to a SAN Volume Controller node or Storwize system node canister.
Details Introduced
During upgrade node failover does not bring up the normal alert message requiring a refresh of the GUI. Customers will need to manually refresh the GUI upon upgrade to v7.8.1.6.

This is a temporary restriction that may be lifted in a future PTF.

7.8.1.6

There is a known issue with 8-node systems and IBM Security Key Lifecycle Manager 3.0 that can cause the status of key server end points, on the system, to occasionally report as degraded or offline. The issue intermittently occurs when the system attempts to validate the key server but the server response times out to some of the nodes. When the issue occurs Error Code 1785 (A problem occurred with the Key Server) will be visible in the system event log.

This issue will not cause any loss of access to encrypted data.

7.8.0.0

There is an extremely small possibility that, on a system using both Encryption and Transparent Cloud Tiering, the system can enter a state where an encryption re-key operation is stuck in 'prepared' or 'prepare_failed' state, and a cloud account is stuck in 'offline' state.

The user will be unable to cancel or commit the encryption rekey, because the cloud account is offline. The user will be unable to remove the cloud account because an encryption rekey is in progress.

The system can only be recovered from this state using a T4 Recovery procedure.

It is also possible that SAS-attached storage arrays go offline.

7.8.0.0

Spectrum Virtualize as Software customers should not enable the Transparent Cloud Tiering function.

This restriction will be removed under APAR HU01495.

7.8.0.0

Due to memory constraints, Storwize V7000 Gen1 control enclosures with one or more compressed volumes cannot be upgraded to v7.8.0.0 or later.

7.8.0.0

Some configuration information will be incorrect in Spectrum Control.

This does not have any functional impact and will be resolved in a future release of Spectrum control.

7.8.0.0

No SAS Direct Attached Windows 2008 Hyper-V hosts will be able to connect to Storwize systems running v7.8.0.0.

7.8.0.0

When a system first upgrades from pre-7.7.0 to 7.7.0 or later, the in-memory audit log will be cleared. This means the catauditlog CLI command, and the GUI, will not show any commands that were issued before the upgrade.

The upgrade test utility will write the existing auditlog to a file in /dumps/audit on the config node. If the auditlog is needed at a later point, this file can be copied from the config node.

Subsequent upgrades will not affect the contents of the auditlog.

7.7.0.0

Priority Flow Control for iSCSI is only supported on Brocade VDX 10GbE switches.

This is a temporary restriction that may be lifted in a future v7.8 PTF.

7.7.0.0

Systems, with NPIV enabled, presenting storage to SUSE Linux Enterprise Server (SLES) or Red Hat Enterprise Linux (RHEL) hosts running the ibmvfc driver on IBM Power can experience path loss or read-only file system events.

This is cause by issues within the ibmvfc driver and VIOS code.

Refer to this troubleshooting page for more information.

n/a
Host Disconnects Using VMware vSphere 5.5.0 Update 2 and vSphere 6.0

Refer to this flash for more information

n/a
If an update stalls or fails then Contact IBM Support for Further Assistance n/a
The following restrictions were valid for previous PTFs, but have now been lifted

In the GUI, when filtering volumes by host, if there are more than 50 host objects, then the host list will not include the host names.

This issue has been resolved by APAR HU01687 in PTF v7.8.1.5.

7.8.1.3

Systems attached to 3PAR controllers cannot be upgraded to v7.8.1.1 or later.

This was a temporary restriction that has been lifted.

7.8.1.1

No upgrade or service activity should be attempted while a Transparent Cloud Tiering Snapshot task is in progress.

This was a temporary restriction that has been lifted.

7.8.0.0

The drive limit remains 1056 drives per cluster.

This was a temporary restriction that has been lifted.

7.8.0.0

3. Issues Resolved

This release contains all of the fixes included in the 7.7.1.3 release, plus the following additional fixes.

A release may contain fixes for security issues, fixes for APARs or both. Consult both tables below to understand the complete set of fixes included in the release.

3.1 Security Issues Resolved

Security issues are documented using a reference number provided by "Common Vulnerabilities and Exposures" (CVE).
CVE Identifier Link for additional Information Resolved in
CVE-2018-25032 6622021 7.8.1.16
CVE-2022-0778 6622017 7.8.1.15
CVE-2021-35603 6622019 7.8.1.15
CVE-2021-35550 6622019 7.8.1.15
CVE-2021-29873 6497111 7.8.1.14
CVE-2020-2781 6445063 7.8.1.13
CVE-2020-13935 6445063 7.8.1.13
CVE-2020-14577 6445063 7.8.1.13
CVE-2020-14578 6445063 7.8.1.13
CVE-2020-14579 6445063 7.8.1.13
CVE-2019-5544 6250889 7.8.1.12
CVE-2019-2964 6250887 7.8.1.12
CVE-2019-2989 6250887 7.8.1.12
CVE-2018-12404 6250885 7.8.1.12
CVE-2019-11477 1164286 7.8.1.11
CVE-2019-11478 1164286 7.8.1.11
CVE-2019-11479 1164286 7.8.1.11
CVE-2019-2602 1073958 7.8.1.10
CVE-2018-3180 ibm10884526 7.8.1.10
CVE-2018-12547 ibm10884526 7.8.1.10
CVE-2008-5161 ibm10874368 7.8.1.9
CVE-2018-5391 ibm10872368 7.8.1.8
CVE-2018-5732 ibm10741135 7.8.1.8
CVE-2018-11776 ibm10741137 7.8.1.8
CVE-2017-17449 ibm10872364 7.8.1.8
CVE-2017-18017 ibm10872364 7.8.1.8
CVE-2018-1517 ibm10872456 7.8.1.8
CVE-2018-2783 ibm10872456 7.8.1.8
CVE-2018-12539 ibm10872456 7.8.1.8
CVE-2018-1775 ibm10872486 7.8.1.8
CVE-2017-17833 ibm10872546 7.8.1.8
CVE-2018-11784 ibm10872550 7.8.1.8
CVE-2016-10708 ibm10717661 7.8.1.6
CVE-2016-10142 ibm10717931 7.8.1.6
CVE-2017-11176 ibm10717931 7.8.1.6
CVE-2018-1433 ssg1S1012263 7.8.1.6
CVE-2018-1434 ssg1S1012263 7.8.1.6
CVE-2018-1438 ssg1S1012263 7.8.1.6
CVE-2018-1461 ssg1S1012263 7.8.1.6
CVE-2018-1462 ssg1S1012263 7.8.1.6
CVE-2018-1463 ssg1S1012263 7.8.1.6
CVE-2018-1464 ssg1S1012263 7.8.1.6
CVE-2018-1465 ssg1S1012263 7.8.1.6
CVE-2018-1466 ssg1S1012263 7.8.1.6
CVE-2016-6210 ssg1S1012276 7.8.1.6
CVE-2016-6515 ssg1S1012276 7.8.1.6
CVE-2013-4312 ssg1S1012277 7.8.1.6
CVE-2015-8374 ssg1S1012277 7.8.1.6
CVE-2015-8543 ssg1S1012277 7.8.1.6
CVE-2015-8746 ssg1S1012277 7.8.1.6
CVE-2015-8812 ssg1S1012277 7.8.1.6
CVE-2015-8844 ssg1S1012277 7.8.1.6
CVE-2015-8845 ssg1S1012277 7.8.1.6
CVE-2015-8956 ssg1S1012277 7.8.1.6
CVE-2016-2053 ssg1S1012277 7.8.1.6
CVE-2016-2069 ssg1S1012277 7.8.1.6
CVE-2016-2384 ssg1S1012277 7.8.1.6
CVE-2016-2847 ssg1S1012277 7.8.1.6
CVE-2016-3070 ssg1S1012277 7.8.1.6
CVE-2016-3156 ssg1S1012277 7.8.1.6
CVE-2016-3699 ssg1S1012277 7.8.1.6
CVE-2016-4569 ssg1S1012277 7.8.1.6
CVE-2016-4578 ssg1S1012277 7.8.1.6
CVE-2016-4581 ssg1S1012277 7.8.1.6
CVE-2016-4794 ssg1S1012277 7.8.1.6
CVE-2016-5412 ssg1S1012277 7.8.1.6
CVE-2016-5828 ssg1S1012277 7.8.1.6
CVE-2016-5829 ssg1S1012277 7.8.1.6
CVE-2016-6136 ssg1S1012277 7.8.1.6
CVE-2016-6198 ssg1S1012277 7.8.1.6
CVE-2016-6327 ssg1S1012277 7.8.1.6
CVE-2016-6480 ssg1S1012277 7.8.1.6
CVE-2016-6828 ssg1S1012277 7.8.1.6
CVE-2016-7117 ssg1S1012277 7.8.1.6
CVE-2016-10229 ssg1S1012277 7.8.1.6
CVE-2016-0634 ssg1S1012278 7.8.1.6
CVE-2017-5647 ssg1S1010892 7.8.1.3
CVE-2016-2183 ssg1S1010205 7.8.1.1
CVE-2016-5546 ssg1S1010205 7.8.1.1
CVE-2016-5547 ssg1S1010205 7.8.1.1
CVE-2016-5548 ssg1S1010205 7.8.1.1
CVE-2016-5549 ssg1S1010205 7.8.1.1
CVE-2017-5638 ssg1S1010113 7.8.1.0
CVE-2016-4461 ssg1S1010883 7.8.1.0
CVE-2016-5385 ssg1S1009581 7.8.0.2
CVE-2016-5386 ssg1S1009581 7.8.0.2
CVE-2016-5387 ssg1S1009581 7.8.0.2
CVE-2016-5388 ssg1S1009581 7.8.0.2
CVE-2016-6796 ssg1S1010114 7.8.0.2
CVE-2016-6816 ssg1S1010114 7.8.0.2
CVE-2016-6817 ssg1S1010114 7.8.0.2
CVE-2016-2177 ssg1S1010115 7.8.0.2
CVE-2016-2178 ssg1S1010115 7.8.0.2
CVE-2016-2183 ssg1S1010115 7.8.0.2
CVE-2016-6302 ssg1S1010115 7.8.0.2
CVE-2016-6304 ssg1S1010115 7.8.0.2
CVE-2016-6306 ssg1S1010115 7.8.0.2
CVE-2016-5696 ssg1S1010116 7.8.0.2
CVE-2016-2834 ssg1S1010117 7.8.0.2
CVE-2016-5285 ssg1S1010117 7.8.0.2
CVE-2016-8635 ssg1S1010117 7.8.0.2
CVE-2017-6056 ssg1S1010022 7.8.0.0

3.2 APARs Resolved

Show details for all APARs
APAR Affected Products Severity Description Resolved in Feature Tags
HU02342 All Critical Occasionally when an offline drive returns to online state later than its peers in the same RAID array there can be multiple node warmstarts that send nodes into a service state (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8 or later
Trigger None
Workaround None
7.8.1.15 RAID
HU02406 All Critical An interoperability issue between Cisco NX-OS firmware and the Spectrum Virtualize Fibre Channel driver can cause a node warmstart on NPIV failback (for example during an upgrade) with the potential for a loss of access. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems using NPIV that are connected to Cisco SAN equipment running NX-OS 8.4(2c) or later
Trigger Initiate an NPIV failback operation by, for example, performing an upgrade
Workaround Disable NPIV (which will require any hot spare nodes to be removed first)
7.8.1.15 Interoperability
HU02471 All Critical After starting a FlashCopy map with -restore in a graph with a GMCV secondary disk that was stopped with -access there can be a data integrity issue (show details)
Symptom Data Integrity Loss
Environment Systems using GMCV
Trigger Given a configuration where there is a GMCV secondary volume A that has a cascade of two FlashCopy maps to volumes B and C (A -> B -> C) The mapping B -> C has been started and GMCV is running Stop the GMCV relationship with -access Start FlashCopy map A -> B with -restore Any I/O to volume A will corrupt data on volume C
Workaround Wait for a GMCV backward map to complete before starting related FlashCopy maps with -restore
7.8.1.15 FlashCopy, Global Mirror with Change Volumes
HU02332 & HU02336 All Suggested When an I/O is received, from a host, with invalid or inconsistent SCSI data but a good checksum it may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8 or later
Trigger None
Workaround None
7.8.1.15 Hosts
HU02429 All Critical System can go offline shortly after changing the SMTP settings using the chemailserver command via the GUI (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8.1 or later
Trigger None
Workaround None
7.8.1.14 System Monitoring
HU02277 All HIPER RAID parity scrubbing can become stalled causing an accumulation of media errors leading to multiple drive failures with the possibility of data integrity loss. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems with model MZILS3T8HMLH read intensive SSDs at drive firmware MS24 are particularly susceptible to the data integrity (DI) issue. Other drive types may see multiple failures without DI issue
Trigger None
Workaround None
7.8.1.13 RAID
HU02338 All HIPER An issue in the setting up of reverse FlashCopy mappings can cause the background copy to finish prematurely providing an incomplete target image (show details)
Symptom Data Integrity Loss
Environment Systems using FlashCopy
Trigger None
Workaround None
7.8.1.13 FlashCopy
HU02222 All Critical Where the source volume of an incremental FlashCopy map is also a Metro or Global Mirror target volume that is using a change volume or is a Hyperswap volume, then there is a possibility that not all data will be copied to the FlashCopy target. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems using Remote Copy
Trigger None
Workaround None
7.8.1.13 Global Mirror with Change Volumes
HU02201 & HU02221 All High Importance Shortly after upgrading drive firmware, specific drive models can fail due to Too many long IOs to drive for too long errors (show details)
Symptom Degraded Operation
Environment Systems with the following drive models:
  • ST300MM0009 (300GB) - B5B8
  • ST600MM0009 (600GB) - B5B8
  • ST900MM0009 (900GB) - B5B8
  • ST1200MM0009 (1200GB) - B5B8
  • ST1200MM0129 (1800GB) - B5C9
  • ST2400MM0129 (2400GB) - B5C9
  • ST300MP0006 (300GB) - B6AA
  • ST600MP0006 (600GB) - B6AA
  • ST900MP0146 (900GB) - B6CB
Trigger None
Workaround None
7.8.1.13 Drives
HU02238 All HIPER Force-stopping a FlashCopy map, where the source volume is a Metro or Global Mirror target volume, may cause other FlashCopy maps to return invalid data if they are not 100% copied, in specific configurations. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems using FlashCopy
Trigger None
Workaround None
7.8.1.12 FlashCopy, Global Mirror, Metro Mirror
HU01970 All Critical When a GMCV relationship is stopped, with the -access option, and the secondary volume is immediately deleted with -force, then all nodes may repeatedly warmstart (show details)
Symptom Loss of Access to Data
Environment Systems using GMCV
Trigger Stop a GMCV relationship with -access and immediately delete the secondary volume
Workaround Do not remove secondary volume, with -force, if the backward FC map from the secondary change volume to the secondary volume is still in progress
7.8.1.12 Global Mirror with Change Volumes
HU02197 All Critical Bulk volume removals can adversely impact related FlashCopy mappings leading to a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems running v7.7.1, or later, using FlashCopy
Trigger None
Workaround None
7.8.1.12 FlashCopy
IT25367 All Critical A T2 recovery may occur when an attempt is made to upgrade, or downgrade, the firmware for an unsupported drive type (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Attempt to upgrade/downgrade the firmware for an unsupported drive type
Workaround None
7.8.1.12 Drives
HU01832 All Suggested Creation and distribution of the config file may cause an out-of-memory condition, leading to a node warmstart (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
7.8.1.12 Reliability Availability Serviceability
HU01868 All Suggested After deleting an encrypted external MDisk, it is possible for the encrypted status of volumes to change to no, even though all remaining MDisks are encrypted (show details)
Symptom None
Environment Systems using encryption
Trigger Delete an encrypted external MDisk
Workaround Ensure that all MDisks in the MDisk group are encrypted - this will ensure that data is encrypted
7.8.1.12 Encryption
HU01917 All Suggested Chrome browser support requires a self-signed certificate to include subject alternate name (show details)
Symptom None
Environment Systems accessed using the Chrome browser
Trigger None
Workaround Accept invalid certificate
7.8.1.12 Graphical User Interface
HU02014 SVC HIPER After a loss of power, where a node has a dead CMOS battery, it will fail to restart correctly. It is possible for both nodes in an I/O group to experience this issue (show details)
Symptom Loss of Access to Data
Environment SVC systems using SV1 model nodes
Trigger None
Workaround None
7.8.1.11 Reliability Availability Serviceability
HU01887 All Critical In circumstances where host configuration data becomes inconsistent, across nodes, an issue in the CLI policing code may cause multiple warmstarts (show details)
Symptom Loss of Access to Data
Environment Systems using Host Clusters
Trigger None
Workaround None
7.8.1.11 Command Line Interface, Host Cluster
HU02043 All Critical Collecting a snap can cause nodes to run out of boot drive space and go offline with node error 565 (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Taking many snap data collections on the same config node
Workaround Manually delete unneeded snaps from the boot drive
7.8.1.11 Support Data Collection
HU02063 All Critical HyperSwap clusters with only two surviving nodes may experience warmstarts on both of those nodes where rcbuffersize is set to 512MB (show details)
Symptom Loss of Access to Data
Environment Systems using Hyperswap
Trigger None
Workaround Reduce rcbuffersize to a value less than 512
7.8.1.11 HyperSwap
IT26257 All Critical Starting a relationship, when the remote volume is offline, may result in a T2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems using Hyperswap
Trigger None
Workaround None
7.8.1.11 HyperSwap
HU01836 All High Importance When an auxiliary volume is moved an issue with pausing the master volume can lead to node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems using Hyperswap
Trigger None
Workaround None
7.8.1.11 HyperSwap
HU01923 All High Importance An issue in the way Global Mirror handles write sequence numbers >512 may cause multiple node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems using GM
Trigger None
Workaround None
7.8.1.11 Global Mirror
HU01944 All High Importance Proactive host failover not waiting for 25 seconds before allowing nodes to go offline during upgrades or maintenance (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
7.8.1.11 Reliability Availability Serviceability
HU01952 SVC, V7000 High Importance When the compression accelerator hardware driver detects an uncorrectable error the node will reboot (show details)
Symptom Loss of Redundancy
Environment Systems using compression cards
Trigger None
Workaround None
7.8.1.11 Compression
HU02049 All High Importance GUI session handling has an issue that can generate many exceptions, adversely impacting GUI performance (show details)
Symptom Performance
Environment All systems
Trigger None
Workaround None
7.8.1.11 Graphical User Interface
HU02149 SVC High Importance When an Enhanced Stretch Cluster is using NPIV, in transitional mode, the path priority is not being reported correctly to some hosts (show details)
Symptom Performance
Environment Systems in an Enhanced Stretch Cluster topology that are using NPIV in its transitional mode
Trigger None
Workaround Manually set the preferred path if possible within the host's MPIO settings. Run NPIV in enabled or disabled mode
7.8.1.11 Hosts
HU01830 All Suggested Missing security-enhancing HTTP response headers (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
7.8.1.11 Security
HU01863 All Suggested In rare circumstances, a drive replacement may result in a ghost drive (i.e. a drive with the same ID as the replaced drive stuck in a permanently offline state) (show details)
Symptom None
Environment All systems
Trigger Drive replacement
Workaround None
7.8.1.11 Drives
HU01892 All Suggested LUNs of greater than 2TB, presented by HP XP7 storage controllers, are not supported (show details)
Symptom Configuration
Environment Systems with HP XP7 backend controllers
Trigger None
Workaround None
7.8.1.11 Backend Storage
HU01981 All Suggested Although an issue, in the HBA firmware, is handled correctly it can still cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
7.8.1.11 Reliability Availability Serviceability
HU01988 All Suggested In the Monitoring -> 3D view page, the "export to csv" button does not function (show details)
Symptom None
Environment Systems running v8.1 or earlier
Trigger None
Workaround None
7.8.1.11 Graphical User Interface
HU02085 All Suggested Freeze time of Global Mirror remote copy consistency groups may not be updated correctly in certain scenarios (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8 or later using Global Mirror
Trigger None
Workaround None
7.8.1.11 Global Mirror
HU02102 All Suggested Excessive processing time required for FlashCopy bitmap operations, associated with large (> 20TB) Global Mirror change volumes, may lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Global Mirror with Change Volumes where some GMCV volumes are >20TB
Trigger None
Workaround Limit GMCV volume capacity to 20TB or less
7.8.1.11 Global Mirror with Change Volumes
HU02126 SVC, V7000, V5000 Suggested There is a low probability that excessive SSH connections may trigger a single node warmstart on the configuration node (show details)
Symptom Single Node Warmstart
Environment Systems with Gen 1 and 2 hardware
Trigger More than one SSH connection attempt per second might occasionally cause the config node to warmstart
Workaround Reduce frequency of SSH connections
7.8.1.11 Command Line Interface
HU01781 All Critical An issue with workload balancing in the kernel scheduler can deprive some processes of the necessary resource to complete successfully resulting in a node warmstarts, that may impact performance, with the possibility of a loss of access to volumes (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround None
7.8.1.10
HU01888 & HU01997 All Critical An issue with restore mappings, in the FlashCopy component, can cause an I/O group to warmstart (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround None
7.8.1.10 FlashCopy
HU01972 All High Importance When an array is in a quiescing state, for example where a member has been deleted, I/O may become pended leading to multiple warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment All systems
Trigger Delete an array member using "charraymember -used unused" command
Workaround None
7.8.1.10 Distributed RAID, RAID
HU00744 All Suggested Single node warmstart due to an accounting issue within the cache component (show details)
Symptom Single Node Warmstart
Environment Systems running v7.3 or later
Trigger None
Workaround None
7.8.1.10 Cache
HU00921 All Suggested A node warmstart may occur when an MDisk state change gives rise to duplicate discovery processes (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
7.8.1.10
HU01737 All Suggested On the Update System screen, for Test Only, if a valid code image is selected, in the Run Update Test Utility dialog, then clicking the Test button will initiate a system update (show details)
Symptom None
Environment All systems
Trigger Select a valid code image in the "Run Update Test Utility" dialog and click "Test" button
Workaround Do not select a valid code image in the "Test utility" field of the "Run Update Test Utility" dialog
7.8.1.10 System Update
HU01915 & IT28654 All Suggested Systems, with encryption enabled, that are using key servers to manage encryption keys, may fail to connect to the key servers if the servers SSL certificates are part of a chain of trust (show details)
Symptom None
Environment Systems with encryption enabled
Trigger None
Workaround None
7.8.1.10 Encryption
HU01617 All HIPER Due to a timing window issue, stopping a FlashCopy mapping, with the -autodelete option, may result in a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems using FlashCopy
Trigger None
Workaround Avoid stopping FlashCopy mappings with the -autodelete option
7.8.1.9 FlashCopy
HU01708 All HIPER A node removal operation during an array rebuild can cause a loss of parity data leading to bad blocks (show details)
Symptom Data Integrity Loss
Environment All systems
Trigger Removing a node during an array rebuild
Workaround Do not remove nodes during an array rebuild
7.8.1.9 RAID
HU01865 All HIPER When creating a HyperSwap relationship, using addvolumecopy (or similar methods), the system should perform a synchronisation operation to copy the data from the original copy to the new copy. In some rare cases this synchronisation is skipped, leaving the new copy with bad data (all zeros) (show details)
Symptom Data Integrity Loss
Environment Systems running v7.5 or later using HyperSwap
Trigger None
Workaround None
7.8.1.9 HyperSwap
HU01913 All HIPER A timing window issue in the DRAID6 rebuild process can cause node warmstarts with the possibility of a loss of access (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID
Trigger None
Workaround None
7.8.1.9 Distributed RAID
HU01723 All Critical A timing window issue, around nodes leaving and re-joining clusters, can lead to hung I/O and node warmstarts (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround None
7.8.1.9 Reliability Availability Serviceability
HU01876 All Critical Where systems are connected to controllers, that have FC ports that are capable of acting as initiators and targets, when NPIV is enabled then node warmstarts can occur (show details)
Symptom Loss of Access to Data
Environment Systems, with NPIV enabled, attached to host ports that can act as SCSI initiators and targets
Trigger Zone host initiator and target ports in with the target port WWPN then enable NPIV
Workaround Unzone host or disable NPIV
7.8.1.9 Backend Storage
IT27460 All Critical Lease expiry can occur between local nodes when remote connection is lost, due to the mishandling of messaging credits (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround Use four ports for local to local node communications, on at least two separate fibre channel adapters per node. Port mask so that all four are usable. Use a different fibre channel adapter than the above two adapters for remote port communications. If there are issues with the FCIP tunnel, temporarily block that until it is fixed.
7.8.1.9 Reliability Availability Serviceability
IT29040 All Critical Occasionally a DRAID rebuild, with drives of 8TB or more, can encounter an issue which causes node warmstarts and potential loss of access (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID with drives of 8TB or more
Trigger None
Workaround None
7.8.1.9 Distributed RAID, RAID
HU01907 SVC High Importance An issue in the handling of the power cable sense registers can cause a node to be put into service state with a 560 error (show details)
Symptom Loss of Redundancy
Environment SVC systems using SV1 model nodes
Trigger None
Workaround None
7.8.1.9 Reliability Availability Serviceability
HU01485 SVC Suggested When a SV1 node is started, with only one PSU powered, powering up the other PSU will not extinguish the Power Fault LED.Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed (show details)
Symptom None
Environment SVC systems using SV1 model nodes
Trigger Power up node with only one PSU powered. Power Fault LED is lit. Power up other PSU. Power Fault LED remains lit.
Workaround Ensure both PSUs are powered before starting node
7.8.1.9 System Monitoring
HU01659 SVC Suggested Node Fault LED can be seen to flash in the absence of an error condition.Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed (show details)
Symptom None
Environment SVC systems using SV1 model nodes
Trigger None
Workaround None
7.8.1.9 System Monitoring
HU01849 All Suggested An excessive number of SSH sessions may lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger Initiate a large number of SSH sessions (e.g. one session every 5 seconds)
Workaround Avoid initiating excessive numbers of SSH sessions
7.8.1.9 System Monitoring
IT26049 All Suggested An issue with CPU scheduling may cause the GUI to respond slowly (show details)
Symptom None
Environment Systems running v7.8 or later
Trigger None
Workaround None
7.8.1.9 Graphical User Interface
HU01492 & HU02024 SVC, V7000, V5000 HIPER All ports of a 16Gb HBA can be affected when a single port is congested. This can lead to lease expiries if all ports used for inter-node communication are on the same FC adapter (show details)
Symptom Loss of Access to Data
Environment Systems using 16Gb HBAs
Trigger All ports used for inter-node communication are on the same FC adapter and a port on that adapter experiences congestion
Workaround Separate inter-node traffic so that multiple adapters are used
7.8.1.8 Reliability Availability Serviceability
HU01726 All HIPER A slow raid member drive in an MDisk may cause node warmstarts and the MDisk to go offline for a short time (show details)
Symptom Offline Volumes
Environment Systems using DRAID
Trigger Failing drives
Workaround None
7.8.1.8 Distributed RAID
HU01940 All HIPER Changing the use of a drive can cause a Tier 2 recovery (warmstarts on all nodes in the cluster). This occurs only if the drive change occurs within a small timing window, so the probability of the issue occurring is low (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Change of drive use
Workaround None
7.8.1.8 Drives
HU01572 All Critical SCSI 3 commands from unconfigured WWPNs may result in multiple warmstarts leading to a loss of access (show details)
Symptom Loss of Access to Data
Environment Systems with iSCSI connected hosts
Trigger None
Workaround None
7.8.1.8 iSCSI
HU01678 All Critical Entering an invalid parameter in the addvdiskaccess command may initiate a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Include an invalid argument in an addvdiskaccess command
Workaround Use care when entering addvdiskaccess CLI commands
7.8.1.8 Command Line Interface
HU01735 All Critical Multiple power failures can cause a RAID array to get into a stuck state leading to offline volumes (show details)
Symptom Offline Volumes
Environment All systems
Trigger Multiple power failures
Workaround None
7.8.1.8 RAID
HU01774 All Critical After a failed mkhost command for an iSCSI host any I/O from that host will cause multiple warmstarts (show details)
Symptom Loss of Access to Data
Environment Systems with iSCSI connected hosts
Trigger mkhost command fails for an iSCSI host
Workaround For iSCSI hosts prevent I/O from the host if the mkhost operation fails
7.8.1.8 iSCSI
HU01799 All Critical Timing window issue can affect operation of the HyperSwap addvolumecopy command causing all nodes to warmstart (show details)
Symptom Loss of Access to Data
Environment Systems using HyperSwap
Trigger None
Workaround Ensuring that the volume is not in a FlashCopy mapping will allow the use of the addvolumecopy. Alternatively the addvolumecopy command combines a number of CLI commands for convenience. Use these CLI commands individually to achieve the same outcome
7.8.1.8 HyperSwap
HU01825 All Critical Invoking a chrcrelationship command when one of the relationships in a consistency group is running in the opposite direction to the others may cause a node warmstart followed by a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems using HyperSwap
Trigger A relationship running in one direction is added to a consistency group running in the other direction whilst one of the FlashCopy maps associated with the HyperSwap relationship is still stopping/cleaning
Workaround Do not add a relationship to a consistency group if they are running in opposite directions (i.e. the Primary of the consistency group and the Primary of the relationship are on different sites); Do not add a relationship to a consistency group if the relationship still has one of its FlashCopy maps in the stopping state. The clean progress needs to reach 100 percent before the relationship can be safely added.
7.8.1.8 FlashCopy
HU01847 All Critical FlashCopy handling of medium errors across a number of drives on backend controllers may lead to multiple node warmstarts (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8.1 or later using FlashCopy
Trigger None
Workaround None
7.8.1.8 FlashCopy
HU01899 All Critical In a HyperSwap cluster, when the primary I/O group has a dead domain, nodes will repeatedly warmstart (show details)
Symptom Loss of Access to Data
Environment Systems using HyperSwap
Trigger A primary volume is offline and remote copy relationship is in a quiesced state
Workaround None
7.8.1.8 HyperSwap
IT25850 All Critical I/O performance may be adversely affected towards the end of DRAID rebuilds. For some systems there may be multiple warmstarts leading to a loss of access (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID
Trigger None
Workaround None
7.8.1.8 Distributed RAID
HU01507 All High Importance Until the initial synchronisation process completes, high system latency may be experienced when a volume is created with two compressed copies or when space-efficient copy is added to a volume with an existing compressed copy (show details)
Symptom Performance
Environment All systems
Trigger Create a volume with two compressed copies or add a space-efficient copy to a volume with an existing compressed copy
Workaround Avoid: creating a new volume with two compressed copies; adding a SE volume copy to a volume that already possesses a compressed copy
7.8.1.8 Volume Mirroring
HU01579 All High Importance In systems where all drives are of type HUSMM80xx0ASS20 it will not be possible to assign a quorum drive (show details)
Symptom Loss of Redundancy
Environment Systems with drives of type HUSMM80xx0ASS20
Trigger Attempt to assign drive type as quorum
Workaround Manually assign a different drive type as quorum
7.8.1.8 Drives, Quorum
HU01661 All High Importance A cache-protection mechanism flag setting can become stuck leading to repeated stops of consistency group synchronisation (show details)
Symptom Loss of Redundancy
Environment Systems running v7.6 or later using remote copy
Trigger None
Workaround None
7.8.1.8 HyperSwap
HU01733 All High Importance Canister information, for the High Density Expansion Enclosure, may be incorrectly reported (show details)
Symptom Loss of Redundancy
Environment Systems using the High Density Expansion Enclosure (92F)
Trigger None
Workaround None
7.8.1.8 Reliability Availability Serviceability
HU01797 All High Importance Hitachi G1500 backend controllers may exhibit higher than expected latency (show details)
Symptom Performance
Environment Systems with Hitachi G1500 backend controllers
Trigger None
Workaround None
7.8.1.8 Backend Storage
HU01813 All High Importance An issue with Global Mirror stream recovery handling at secondary sites can adversely impact replication performance (show details)
Symptom Performance
Environment Systems using Global Mirror
Trigger Secondary cluster overloaded
Workaround None
7.8.1.8 Global Mirror
HU01824 All High Importance Switching replication direction for HyperSwap relationships can lead to long I/O timeouts (show details)
Symptom Performance
Environment Systems using HyperSwap
Trigger Switch replication direction of a HyperSwap relationship
Workaround None
7.8.1.8 HyperSwap
HU01839 All High Importance Where a VMware host is being served volumes, from two different controllers, and an issue, on one controller, causes the related volumes to be taken offline then I/O performance, for the volumes from the other controller, will be adversely affected (show details)
Symptom Performance
Environment Systems running v7.5 or later presenting volumes to VMware hosts, from more than one back-end controller
Trigger Issue on back-end controller takes volumes offline
Workaround None
7.8.1.8 Hosts
HU01842 All High Importance Bursts of I/O to Read-Intensive Drives can be interpreted as dropped frames against the resident slots, leading to redundant drives being incorrectly failed (show details)
Symptom Loss of Redundancy
Environment Systems with Read-Intensive Drives
Trigger None
Workaround None
7.8.1.8 Drives
HU01846 SVC High Importance Silent battery discharge condition will unexpectedly take a SVC node offline putting it into a 572 service state (show details)
Symptom Loss of Redundancy
Environment SVC systems using DH8 & SV1 model nodes
Trigger None
Workaround None
7.8.1.8 Reliability Availability Serviceability
HU01902 V7000, V5000, V3700, V3500 High Importance During an upgrade, an issue with VPD migration, can cause a timeout leading to a stalled upgrade (show details)
Symptom Loss of Redundancy
Environment Storwize systems
Trigger Upgrade
Workaround None
7.8.1.8 System Update
HU01276 All Suggested An issue in the handling of debug data from the FC adapter can cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
7.8.1.8 Reliability Availability Serviceability
HU01467 All Suggested Failures in the handling of performance statistics files may lead to missing samples in Spectrum Control and other tools (show details)
Symptom None
Environment All systems
Trigger None
Workaround Increase the sampling interval
7.8.1.8 System Monitoring
HU01512 All Suggested During a DRAID MDisk copy-back operation a miscalculation of the remaining work may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using DRAID
Trigger None
Workaround None
7.8.1.8 Distributed RAID
HU01523 All Suggested An issue with FC adapter initialisation can lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
7.8.1.8 Reliability Availability Serviceability
HU01556 All Suggested The handling of memory pool usage by Remote Copy may lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.3 or later
Trigger None
Workaround None
7.8.1.8 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01564 All Suggested FlashCopy maps cleaning process is not monitoring the grains correctly which may cause FlashCopy maps to not stop (show details)
Symptom None
Environment Systems using FlashCopy
Trigger None
Workaround None
7.8.1.8 FlashCopy
HU01657 SVC, V7000, V5000 Suggested The 16Gb FC HBA firmware may experience an issue, with the detection of unresponsive links, leading to a single node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
7.8.1.8 Reliability Availability Serviceability
HU01715 All Suggested Issuing a rmvolumecopy command followed by an expandvdisksize command may result in hung I/O leading to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using HyperSwap
Trigger A rmvolumecopy command followed by an expandvdisksize command
Workaround Stop I/O to the volume approx 10 mins prior to expanding it
7.8.1.8 HyperSwap
HU01719 All Suggested Node warmstart due to a parity error in the HBA driver firmware (show details)
Symptom Single Node Warmstart
Environment Systems running v7.6 and later using 16Gb HBAs
Trigger None
Workaround None
7.8.1.8 Reliability Availability Serviceability
HU01751 All Suggested When RAID attempts to flag a strip as bad, and that strip has already been flagged, a node may warmstart (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
7.8.1.8 RAID
HU01760 All Suggested FlashCopy map progress appears to be stuck at zero percent (show details)
Symptom None
Environment Systems using FlashCopy
Trigger None
Workaround None
7.8.1.8 FlashCopy
HU01786 All Suggested An issue in the monitoring of SSD write endurance can result in false 1215/2560 errors in the Event Log (show details)
Symptom None
Environment Systems running v7.7.1 or later with SSDs
Trigger None
Workaround None
7.8.1.8 Drives
HU01790 All Suggested On the Create Volumes page the Accessible I/O Groups selection may not update when the Caching I/O group selection is changed (show details)
Symptom None
Environment Systems with more than one I/O group
Trigger Change "Caching I/O group" selection on "Create Volumes"
Workaround Leave the "Caching I/O group" and "Accessible I/O Groups" selections as "default". Use "modify I/O group" action (right-click "volume" -> "modify I/O group"") to modify volume's iogrp.
7.8.1.8 Graphical User Interface
HU01793 All Suggested The Maximum final size value in the Expand Volume dialog can display an incorrect value preventing expansion (show details)
Symptom None
Environment All systems
Trigger Set volume mirroring memory to minimum. Create a generic mirrored volume of 2TB or less. Open Expand Volume dialog
Workaround Use CLI to expand volume
7.8.1.8 Graphical User Interface
HU02028 All Suggested An issue, with timer cancellation, in the Remote Copy component may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Remote Copy
Trigger None
Workaround None
7.8.1.8 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
IT19561 All Suggested An issue with register clearance in the FC driver code may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
7.8.1.8 Reliability Availability Serviceability
IT22591 All Suggested An issue in the HBA adapter firmware may result in node warmstarts (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
7.8.1.8 Reliability Availability Serviceability
IT24900 V7000, V5000 Suggested Whilst replacing a control enclosure midplane an issue at boot can prevent VPD being assigned delaying a return to service (show details)
Symptom None
Environment All Storwize systems
Trigger None
Workaround None
7.8.1.8 Reliability Availability Serviceability
IT26836 V7000, V5000 Suggested Loading drive firmware may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Storwize systems
Trigger Loading drive firmware
Workaround None
7.8.1.8 Drives
HU01802 All Critical USB encryption key can become inaccessible after upgrade. If the system is later rebooted then any encrypted volumes will be unavailable (show details)
Symptom Loss of Access to Data
Environment Systems using encryption
Trigger System Upgrade
Workaround None
7.8.1.7 Encryption
HU01785 All High Importance An issue with memory mapping may lead to multiple node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment All systems
Trigger None
Workaround None
7.8.1.7
HU01792 All HIPER When a DRAID array has multiple drive failures and the number of failed drives is greater than the number of rebuild areas in the array it is possible that the storage pool will be taken offline during the copyback phase of a rebuild. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID
Trigger None
Workaround None
7.8.1.6 Distributed RAID
HU01866 SVC HIPER A faulty PSU sensor, in a node, can fill the sel log causing the service processor (BMC) to disable logging. If a snap is subsequently taken, from the node, a timeout will occur and it will be taken offline. It is possible for this to affect both nodes in an I/O group (show details)
Symptom Loss of Access to Data
Environment SVC systems using SV1 model nodes
Trigger None
Workaround None
7.8.1.6 System Monitoring
HU01524 All Critical When a system loses input power, nodes will shut down until power is restored. If a node was in the process of creating a bad block for an MDisk, at the moment it shuts down, then there is a chance that the system will hit repeated Tier 2 recoveries when it powers back up (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8.1 or later
Trigger None
Workaround None
7.8.1.6 RAID
HU01767 All Critical Reads of 4K/8K from an array can under exceptional circumstances return invalid data. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8.0 or earlier
Trigger None
Workaround None
7.8.1.6 RAID, Thin Provisioning
IT17919 All Critical A rare timing window issue in the handling of Remote Copy state can result in multi-node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v7.2 or later using Remote Copy
Trigger None
Workaround None
7.8.1.6 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01420 All High Importance An issue in DRAID can cause repeated node warmstarts in the circumstances of a degraded copyback operation to a drive (show details)
Symptom Multiple Node Warmstarts
Environment Systems using DRAID
Trigger None
Workaround None
7.8.1.6 Distributed RAID
HU01476 All High Importance A remote copy relationship may suffer a loss of synchronisation when the relationship is renamed (show details)
Symptom Loss of Redundancy
Environment Systems using Remote Copy
Trigger Rename remote copy relationship using GUI or CLI
Workaround None
7.8.1.6 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01623 All High Importance An issue in the handling of inter-node communications can lead to latency for Remote Copy relationships (show details)
Symptom Performance
Environment Systems running v7.1 or later using Remote Copy
Trigger None
Workaround None
7.8.1.6 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01630 All High Importance When a system with FlashCopy mappings is upgraded there may be multiple node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v7.5 or earlier using FlashCopy
Trigger Upgrade to v7.6 or later
Workaround None
7.8.1.6 FlashCopy
HU01697 All High Importance A timeout issue in RAID member management can lead to multiple node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment All systems
Trigger None
Workaround None
7.8.1.6 RAID
HU01771 SVC, V7000 High Importance An issue with the CMOS battery in a node can cause an unexpectedly large log file to be generated by the BMC. At log collection the node may be taken offline (show details)
Symptom Loss of Redundancy
Environment SVC & V7000 systems running v7.8 or later
Trigger Node CMOS battery issue
Workaround None
7.8.1.6 System Monitoring
HU01446 All Suggested Where host workload overloads the back-end controller and VMware hosts are issuing ATS commands a race condition may be triggered leading to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.6.1 or later with VMware hosts using VAAI CAW feature
Trigger None
Workaround Avoid overloading the back-end
7.8.1.6 Hosts
HU01472 All Suggested A locking issue in Global Mirror can cause a warmstart on the secondary cluster (show details)
Symptom Single Node Warmstart
Environment Systems using Global Mirror
Trigger None
Workaround None
7.8.1.6 Global Mirror
HU01619 All Suggested A misreading of the PSU register can lead to failure events being logged incorrectly (show details)
Symptom None
Environment Systems running v7.6 or later
Trigger None
Workaround None
7.8.1.6 System Monitoring
HU01628 All Suggested In the GUI on the Volumes page whilst using the filter function some volumes entries may not be displayed until the page has completed loading (show details)
Symptom None
Environment All systems
Trigger Enter a filter string on the Volumes page
Workaround None
7.8.1.6 Graphical User Interface
HU01664 All Suggested A timing window issue during an upgrade can cause the node restarting to warmstart stalling the upgrade (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8 or later
Trigger None
Workaround None
7.8.1.6 System Update
HU01698 All Suggested A node warmstart may occur when deleting a compressed volume if a host has written to the volume minutes before the volume is deleted (show details)
Symptom Single Node Warmstart
Environment Systems with compressed volumes
Trigger Write I/O received by compression code for a volume that has been deleted
Workaround Remove volume host mappings and leave an interval of time (90mins) before removing the volume
7.8.1.6 Compression
HU01740 All Suggested The timeout setting for key server commands may be too brief, when the server is busy, causing those commands to fail (show details)
Symptom None
Environment Systems running v7.8 or later using encryption
Trigger Enter mkkeyserver command
Workaround Retry command
7.8.1.6 Encryption
HU01747 All Suggested The incorrect detection of a cache issue can lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8 or later
Trigger None
Workaround None
7.8.1.6 Cache
HU00247 All Critical A rare deadlock condition can lead to a RAID5 or RAID6 array rebuild stalling at 99% (show details)
Symptom Loss of Access to Data
Environment Systems using RAID5/6
Trigger None
Workaround None
7.8.1.5 Distributed RAID, RAID
HU01620 All Critical Configuration changes can slow critical processes and, if this coincides with cloud account statistical data being adjusted, a Tier 2 recovery may occur (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8 or later that support TCT
Trigger Configuration changes around midnight
Workaround Avoid making large numbers of configuration changes around midnight
7.8.1.5 Transparent Cloud Tiering
IC57642 All Critical A complex combination of failure conditions in the fabric connecting nodes can result in lease expiries, possibly cluster-wide (show details)
Symptom Loss of Access to Data
Environment Systems with more than 2 nodes
Trigger None
Workaround None
7.8.1.5 Reliability Availability Serviceability
IT19192 All Critical An issue in the handling of GUI certificates may cause warmstarts leading to a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround None
7.8.1.5 Graphical User Interface, Reliability Availability Serviceability
IT23747 All High Importance For large drive sizes the DRAID rebuild process can consume significant CPU resource adversely impacting system performance (show details)
Symptom Performance
Environment Systems running v7.6 or later using DRAID
Trigger High write workloads during a DRAID rebuild
Workaround Reduce write workload during rebuild or use a volume/pool level throttle
7.8.1.5 Distributed RAID
HU01655 All Suggested The algorithm used to calculate an SSDs replacement date can sometimes produce incorrect results leading to a premature End-of-Life error being reported (show details)
Symptom None
Environment Systems using SSDs
Trigger None
Workaround Ignore error if SSD write endurance (in lsdrive CLI output) shows less than 80
7.8.1.5 Drives
HU01679 All Suggested An issue in the RAID component can very occasionally cause a single node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.7 or later
Trigger None
Workaround None
7.8.1.5 RAID
HU01687 All Suggested For volumes by host, ports by host and volumes by pool pages in the GUI when the number of items is greater than 50 then the item name will not be displayed (show details)
Symptom None
Environment Systems running v7.7 or v7.8
Trigger More than 50 hosts/pools
Workaround Use CLI
7.8.1.5 Graphical User Interface
HU01704 All Suggested In systems using HyperSwap a rare timing window issue can result in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using HyperSwap
Trigger None
Workaround None
7.8.1.5 HyperSwap
HU01724 All Suggested An I/O lock handling issue between nodes can lead to a single node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8 or later
Trigger None
Workaround None
7.8.1.5 RAID
HU01729 All Suggested Remote copy uses multiple streams to send data between clusters. During a stream disconnect a node, unable to progress, may warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Remote Copy
Trigger None
Workaround None
7.8.1.5 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01730 SVC Suggested When running the DMP for a 1046 error the picture may not indicate the correct position of the failed adapter (show details)
Symptom None
Environment DH8 systems
Trigger Run DMP for 1046 error
Workaround Be aware of which adapter is failed and do not use the picture in the GUI
7.8.1.5 GUI Fix Procedure
HU01731 SVC, V7000, V5000 Suggested When a node is placed into service mode it is possible for all compression cards within the node to be marked as failed (show details)
Symptom None
Environment Systems running v7.7 or later with compression cards
Trigger Put node in service state
Workaround None
7.8.1.5 Compression
HU01763 SVC Suggested A single node warmstart may occur on a DH8 config node when inventory email is created. The issue only occurs if this coincides with a very high rate of CLI commands and high I/O workload on the config node (show details)
Symptom Single Node Warmstart
Environment DH8 systems running v7.7.1 or later
Trigger High CLI workload and I/O workload at same time as inventory email
Workaround None
7.8.1.5 Command Line Interface, System Monitoring
IT23140 All Suggested When viewing the licensed functions GUI page the individual calculations for SCUs, for each tier, may be wrong. However the total is correct (show details)
Symptom None
Environment All systems
Trigger None
Workaround Use CLI
7.8.1.5 Graphical User Interface
HU01706 All HIPER Areas of volumes written with all-zero data may contain non-zero data. For more details refer to this Flash (show details)
Symptom Incorrect data read from volume
Environment Systems running 7.7.1.7 or 7.8.1.3
Trigger See Flash
Workaround None
7.8.1.4
HU01646 All HIPER A new failure mechanism in the 16Gb HBA driver can under certain circumstances lead to a lease expiry of the entire cluster (show details)
Symptom Loss of Access to Data
Environment Systems with 16Gb HBAs
Trigger Faulty SAN hardware (adapter; SFP; switch)
Workaround None
7.8.1.3 Reliability Availability Serviceability
HU01321 All Critical Multi-node warmstarts may occur when changing the direction of a remote copy relationship whilst write I/O to the (former) primary volume is still occurring (show details)
Symptom Multiple Node Warmstarts
Environment Systems using Remote Copy
Trigger Change the direction of a remote copy relationship
Workaround None
7.8.1.3 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01490 All Critical When attempting to add/remove multiple IQNs to/from a host the tables that record host-wwpn mappings can become inconsistent resulting in repeated node warmstarts across I/O groups (show details)
Symptom Loss of Access to Data
Environment Systems with iSCSI connected hosts
Trigger
  • A addhostport command with iqn2 and iqn1 (where iqn1 is already recorded) is entered;
  • This command attempts to add iqn2 but determines that iqn1 is a duplicate and the CLI command fails;
  • Later whenever a login request from iqn2 is received internal checking detects an inconsistency and warmstarts the node
Workaround Do not use multiple IQNs in iSCSI add/remove commands
7.8.1.3 iSCSI
HU01549 All Critical During a system upgrade HyperV-clustered hosts may experience a loss of access to any iSCSI connected volumes (show details)
Symptom Loss of Access to Data
Environment Systems running v7.5 or earlier with iSCSI connected HyperV clustered hosts
Trigger Upgrade to v7.6 or later
Workaround None
7.8.1.3 System Update, iSCSI
HU01625 All Critical In systems with a consistency group of HyperSwap or Metro Mirror relationships if an upgrade attempts to commit whilst a relationship is out of synch then there may be multiple warmstarts and a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems running v7.7 or earlier using HyperSwap or Metro Mirror
Trigger Upgrading to v7.8 or later
Workaround Stop all replication prior to upgrading
7.8.1.3 System Update, HyperSwap, Metro Mirror
IT23034 All Critical With HyperSwap volumes and mirrored copies, at a single site, using rmvolumecopy to remove a copy, from an auxiliary volume, may result in a cluster-wide warmstart necessitating a Tier 2 recovery (show details)
Symptom Multiple Node Warmstarts
Environment Systems with HyperSwap volumes and mirrored copies at a single site
Trigger Using rmvolumecopy to remove a copy from a HyperSwap auxiliary volume
Workaround None
7.8.1.3 HyperSwap
HU01481 All High Importance A failed I/O can trigger HyperSwap to unexpectedly change the direction of the relationship leading to node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v7.5 or later using HyperSwap
Trigger None
Workaround None
7.8.1.3 HyperSwap
HU01525 All High Importance During an upgrade a resource locking issue in the compression component can cause a node to warmstart multiple times and become unavailable (show details)
Symptom Loss of Redundancy
Environment Systems running v7.6 or earlier using compressed volumes
Trigger Upgrade to v7.7 or later
Workaround None
7.8.1.3 Compression, System Update
HU01569 SVC High Importance When compression utilisation is high the config node may exhibit longer I/O response times than non-config nodes (show details)
Symptom Performance
Environment SVC systems using compression
Trigger High compression workloads
Workaround None
7.8.1.3 Compression
HU01584 All High Importance An issue in array indexing can cause a RAID array to go offline repeatedly (show details)
Symptom Offline Volumes
Environment System running v7.6 or later
Trigger None
Workaround Avoid doing member exchanges
7.8.1.3 RAID
HU01614 All High Importance After a node is upgraded hosts defined as TPGS may have paths set to inactive (show details)
Symptom Loss of Redundancy
Environment Systems running v7.6 or earlier with host type TPGS
Trigger Upgrade to v7.7 or later
Workaround None
7.8.1.3 Hosts
HU01632 All High Importance A congested fabric causes the Fibre Channel adapter firmware to abort I/O resulting in node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment All systems
Trigger Congested fabric
Workaround Avoid fabric congestion
7.8.1.3 Reliability Availability Serviceability
HU01636 V5000, V3700, V3500 High Importance A connectivity issue with certain host SAS HBAs can prevent hosts from establishing stable communication with the storage controller (show details)
Symptom Performance
Environment Systems presenting storage to hosts with N2225 adapters
Trigger Host with N2225 adapters running Windows Server 2012R2
Workaround None
7.8.1.3 Hosts
HU01638 All High Importance When upgrading to v7.6 or later, if there is another cluster in the same zone which is at v5.1 or earlier then nodes will warmstart and the upgrade will fail (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v7.5 or earlier
Trigger Upgrade to v7.6 or later with a cluster in the same zone running v6.1 or earlier
Workaround Unzone any cluster running v5.1 or earlier from the cluster being upgraded
7.8.1.3 System Update
HU01645 SVC High Importance After upgrading to v7.8 a reboot of a node will initiate a continual boot cycle (show details)
Symptom Loss of Redundancy
Environment DH8 systems running v7.7.0 or earlier with 32GB of RAM
Trigger Upgrade to v7.8; Reboot node
Workaround None
7.8.1.3 System Update
HU01385 All Suggested A warmstart may occur if a rmvolumecopy or rmrcrelationship command are issued on a volume while I/O is being forwarded to the associated copy (show details)
Symptom Single Node Warmstart
Environment Systems running v7.6 or v7.7 that are using Hyperswap
Trigger Issue an rmvolumecopy or rmrcrelationship command whilst hosts are still actively using the Hyperswap volume
Workaround Do not remove a Hyperswap volume or relationship whilst hosts are still mapped to it
7.8.1.3 HyperSwap
HU01457 V7000 Suggested In a hybrid V7000 cluster where one I/O group supports 10k volumes and another does not some operations on volumes may incorrectly be denied in the GUI (show details)
Symptom None
Environment Systems running v7.7.1 or later
Trigger None
Workaround Performed required actions in the CLI
7.8.1.3 Graphical User Interface
HU01535 All Suggested An issue with Fibre Channel driver handling of command processing can result in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.6 or later using Fibre Channel connectivity
Trigger None
Workaround None
7.8.1.3
HU01563 V7000 Suggested Where an IBM SONAS host id is used it can under rare circumstances cause a warmstart (show details)
Symptom Single Node Warmstart
Environment Unified configurations
Trigger None
Workaround None
7.8.1.3
HU01582 All Suggested A compression issue in IP replication can result in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.7 or later using IP Replication
Trigger None
Workaround None
7.8.1.3 IP Replication
HU01624 All Suggested GUI response can become very slow in systems with a large number of compressed and uncompressed volume (show details)
Symptom None
Environment Systems that have a large number of volumes (1000+) including some that are compressed
Trigger None
Workaround None
7.8.1.3 Graphical User Interface
HU01631 All Suggested A memory leak in EasyTier when pools are in Balanced mode can lead to node warmstarts (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8 or later using EasyTier
Trigger Pools in Balanced mode
Workaround Create hybrid pools with multiple tiers of storage and set Easy tier mode to Active
7.8.1.3 EasyTier
HU01654 SVC, V7000, V5000 Suggested There may be a node warmstart when a switch of direction, in a HyperSwap relationship, fails to complete properly (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8 or later using HyperSwap
Trigger None
Workaround None
7.8.1.3 HyperSwap
HU01239 & HU01255 & HU01586 All HIPER The presence of a faulty SAN component can delay lease messages between nodes leading to a cluster-wide lease expiry and consequential loss of access (show details)
Symptom Loss of Access to Data
Environment Systems with 16Gb HBAs
Trigger Faulty SAN hardware (adapter, SFP, switch)
Workaround None
7.8.1.2 Reliability Availability Serviceability
HU01626 All High Importance Node downgrade from v7.8.x to v7.7.1 or earlier (e.g. during an aborted upgrade) may prevent the node from rejoining the cluster. Systems that have already completed upgrade to v7.8.x are not affected by this issue (show details)
Symptom Loss of Redundancy
Environment Systems running v7.8 or later
Trigger Downgrade to v7.7.1 or earlier
Workaround None
7.8.1.2 System Update
HU01505 All HIPER A non-redundant drive experiencing many errors can be taken offline obstructing rebuild activity (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Drive with many errors
Workaround Bring drive online
7.8.1.1 Backend Storage, RAID
HU01570 V7000, V5000, V3700, V3500 HIPER Reseating a drive in an array may cause the MDisk to go offline (show details)
Symptom Offline Volumes
Environment Storwize Gen 1 systems running v7.8 or later
Trigger Removing a drive at the point of failure
Workaround Only remove a drive when offline/failed
7.8.1.1 Drives, RAID
IT20627 All Critical When Read-Intensive drives are used as quorum disks a drive outage can occur. Under some circumstances this can lead to a loss of access (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Using Read-Intensive drives as quorum disks
Workaround None
7.8.1.1 Quorum
HU01477 V7000, V5000, V3700, V3500 High Importance Due to the way enclosure data is read it is possible for a firmware mismatch between nodes to occur during an upgrade (show details)
Symptom Loss of Redundancy
Environment Storwize systems running v7.8 or later
Trigger Upgrade from v7.8.0 to v7.8.1
Workaround Check for dependent volumes. If there are none then manually load enclosure firmware from service mode
7.8.1.1 System Update
HU01503 All High Importance When the 3PAR host type is set to legacy the round robin algorithm, used to select the MDisk port for I/O submission to 3PAR controllers, does not work correctly and I/O may be submitted to fewer controller ports, adversely affecting performance (show details)
Symptom Performance
Environment Systems running v7.5.0.10 v7.6.1.6 or v7.7.1.5 that are virtualising 3PAR storage subsystems
Trigger Host persona set to 6 (legacy) on 3PAR controller
Workaround Change 3PAR host persona to 2 (ALUA) instead of 6 (legacy)
7.8.1.1 Backend Storage
HU01609 & IT15343 All High Importance When the system is busy, the compression component may be paged out of memory resulting in latency that can lead to warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems using compression
Trigger None
Workaround Reduce compression workload
7.8.1.1 Compression
IT19726 SVC High Importance Warmstarts may occur when the attached SAN fabric is congested and HBA transmit paths become stalled preventing the HBA firmware from generating the completion for a FC command (show details)
Symptom Multiple Node Warmstarts
Environment SVC systems
Trigger SAN congestion
Workaround None
7.8.1.1 Hosts
HU00763 & HU01237 V7000, V5000, V3700, V3500 Suggested A node warmstart may occur when a quorum disk is accessed at the same time as the login to that disk is closed (show details)
Symptom Single Node Warmstart
Environment Storwize systems
Trigger None
Workaround None
7.8.1.1 Quorum
HU01332 All Suggested Performance monitor and Spectrum Control show zero CPU utilisation for compression (show details)
Symptom None
Environment Systems running v7.6 or later
Trigger None
Workaround None
7.8.1.1 System Monitoring
HU01353 All Suggested CLI allows the input of carriage return characters into certain fields, after cluster creation, resulting in invalid cluster VPD and failed node adds (show details)
Symptom Configuration
Environment All systems
Trigger After cluster creation use the CLI to enter a carriage return in a command that allows free text in an argument
Workaround Do not insert a carriage return character into text being entered via CLI
7.8.1.1 Command Line Interface
HU01391 & HU01581 V7000, V5000, V3700, V3500 Suggested Storwize systems may experience a warmstart due to an uncorrectable error in the SAS firmware (show details)
Symptom Single Node Warmstart
Environment Storwize systems
Trigger None
Workaround None
7.8.1.1 Drives
HU01430 V7000, V5000, V3700, V3500 Suggested Memory resource shortages in systems with 8GB of RAM can lead to node warmstarts (show details)
Symptom Single Node Warmstart
Environment Storwize Gen 1 systems running v7.6 or later
Trigger None
Workaround None
7.8.1.1
HU01469 V3700, V3500 Suggested Resource exhaustion in the iSCSI component can result in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Storwize V3k systems using iSCSI
Trigger None
Workaround None
7.8.1.1 iSCSI
HU01471 V5000 Suggested Power system down using the GUI on V5000 causes the fans to run high while the system is offline but power is still applied to the enclosure (show details)
Symptom None
Environment Storwize V5k systems running v7.6 or later
Trigger svctask stopsystem -gui
Workaround After shutdown remove power from enclosure
7.8.1.1
HU01484 All Suggested During a RAID array rebuild there may be node warmstarts (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
7.8.1.1 RAID
HU01496 SVC Suggested SVC node type SV1 reports wrong FRU part number for compression accelerator (show details)
Symptom None
Environment SVC SV1 systems
Trigger Run lsnodevpd
Workaround If a replacement is required order FRU part number 01EJ817
7.8.1.1 Command Line Interface
HU01520 V3700, V3500 Suggested Where the system is being used as secondary site for Remote Copy during an upgrade to v7.8.1 the node may warmstart (show details)
Symptom Single Node Warmstart
Environment Storwize V3500 & V3700 systems running v7.1 or later being used as a secondary site
Trigger Upgrade to v7.8.1
Workaround Stop all replication prior to an upgrade to v7.8.1
7.8.1.1 System Update, Global Mirror, Metro Mirror
HU01531 All Suggested Spectrum Control is unable to receive notifications from SVC/Storwize. Spectrum Control may experience an out-of-memory condition (show details)
Symptom None
Environment Systems running v7.8.1.0 that are being managed by IBM Spectrum Control
Trigger None
Workaround None
7.8.1.1 System Monitoring
HU01566 SVC Suggested After upgrading, numerous 1370 errors are seen in the Event Log (show details)
Symptom None
Environment SVC CG8 systems containing old SSD drives (type Z16IZD2B-146)
Trigger System Upgrade
Workaround Ignore 1370 errors
7.8.1.1 System Update
IT19973 All Suggested Call home emails may not be sent due to a failure to retry (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
7.8.1.1
HU01474 SVC, V7000 HIPER Host writes to a read-only secondary volume trigger I/O timeout warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment SVC SV1 & V7000 Gen 2+ systems running v7.7 or later using Remote Copy
Trigger Host write I/O to a read-only secondary volume
Workaround Do not attempt to issue host writes to a secondary volume while remote copy is active
7.8.1.0 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01479 All HIPER The handling of drive reseats can sometimes allow I/O to occur before the drive has been correctly failed resulting in offline MDisks (show details)
Symptom Loss of Access to Data
Environment Systems running v7.6 or later using DRAID
Trigger Reseat a failed drive
Workaround Rather than reseating the drive use the CLI or GUI to fail then unfail it
7.8.1.0 Distributed RAID
HU01483 All HIPER mkdistributedarray command may get stuck in the prepare state. Any interaction with the volumes in that array will result in multiple warmstarts (show details)
Symptom Loss of Access to Data
Environment Systems running v7.7.1 or later
Trigger mkdistributedarray command gets stuck in the prepare state; run rmarray or rmvdisk commands or generate I/O to volumes created on that array
Workaround None
7.8.1.0 Distributed RAID
HU01675 V7000 HIPER Memory allocation issues may cause GUI and I/O performance issues (show details)
Symptom Loss of Access to Data
Environment V7000 Gen2/2+ systems running v7.7.1.7 or later using compression
Trigger Upgrade to v7.7.1.7 or later
Workaround None
7.8.1.0 Compression, Graphical User Interface
HU01220 All Critical Changing the type of a RC consistency group when a volume in a subordinate relationship is offline will cause a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems running v7.5 or later using GMCV
Trigger GMCV relationship with offline volume; Change overarching consistency group type from GMCV to GM
Workaround Identify relationships with offline change volumes and remove them from the consistency group before changing the group's type. Change the type of the relationship separately and then add it back into the group.
7.8.1.0 Global Mirror with Change Volumes, Global Mirror
HU01252 SVC Critical Where a SVC is presenting storage from an 8-node V7000, an upgrade to that V7000 can pause I/O long enough for the SVC to take related MDisks offline (show details)
Symptom Offline Volumes
Environment SVC systems with large and complex V7000 configurations
Trigger None
Workaround None
7.8.1.0
HU01416 All Critical ISL configuration activity may cause a cluster-wide lease expiry (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround None
7.8.1.0 Reliability Availability Serviceability
HU00747 V7000, V5000, V3700, V3500 High Importance Node warmstarts can occur when drives become degraded (show details)
Symptom Multiple Node Warmstarts
Environment Storwize systems running v7.3 or later
Trigger Degraded drives
Workaround Avoid degraded drives
7.8.1.0 Backend Storage
HU01309 SVC High Importance For FC logins, on a node that is online for more than 200 days, if a fabric event makes a login inactive then the node may be unable to re-establish the login (show details)
Symptom Loss of Redundancy
Environment Systems with 8GB FC cards
Trigger FC login established; node up for >200 days; fabric event causes login to go inactive
Workaround None
7.8.1.0 Backend Storage
HU01371 SVC, V7000, V5000 High Importance A remote copy command related to HyperSwap may hang resulting in a warmstart of the config node (show details)
Symptom Loss of Redundancy
Environment Systems running v7.5 or later using HyperSwap
Trigger None
Workaround None
7.8.1.0 HyperSwap
HU01388 SVC, V7000, V5000 High Importance Where a HyperSwap volume is the source of a FlashCopy mapping and the HyperSwap relationship is out of sync when the HyperSwap volume comes back online a switch of direction will occur and FlashCopy operation may delay I/O leading to node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems using FlashCopy with HyperSwap
Trigger An offline HyperSwap volume that is the source of a FlashCopy mapping is brought back online
Workaround Stop dependent FlashCopy mappings or let them complete before bringing an offline HyperSwap volume back online
7.8.1.0 HyperSwap, FlashCopy
HU01394 All High Importance Node warmstarts may occur on systems which are performing Global Mirror replication, due to a low-probability timing window (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v7.2 or later using GM
Trigger None
Workaround None
7.8.1.0 Global Mirror
HU01395 All High Importance Malformed URLs sent by security scanners whilst correctly discarded can cause considerable exception logging on config nodes leading to performance degradation that can adversely affect remote copy (show details)
Symptom Performance
Environment Systems using GM
Trigger Security scanning config node
Workaround Avoid security scanning Storwize clusters
7.8.1.0 Global Mirror
HU01413 All High Importance Node warmstarts when establishing an FC partnership between a system on v7.7.1 or later with another system which in turn has a partnership to another system running v6.4.1 or earlier (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v7.7.1 or later with partnerships to systems which have partnerships with systems running v6.4.1 or earlier
Trigger Create FC partnership
Workaround Upgrade systems on pre-7.2 code levels
7.8.1.0 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01428 V7000, V5000, V3700, V3500 High Importance Scheduling issue adversely affects performance resulting in node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Storwize systems running v7.7.1 or later
Trigger None
Workaround None
7.8.1.0 Reliability Availability Serviceability
HU01480 All High Importance Under some circumstances the config node does not fail over properly when using IPv6 adversely affecting management access via GUI and CLI (show details)
Symptom Configuration
Environment Systems using IPv6 cluster addresses
Trigger None
Workaround None
7.8.1.0 Command Line Interface, Graphical User Interface
IT19019 V5000 High Importance V5000 control enclosure midplane FRU replacement may fail leading to both nodes reporting a 506 error (show details)
Symptom None
Environment V5000 systems running v7.8 or later
Trigger None
Workaround None
7.8.1.0 Reliability Availability Serviceability
HU01057 SVC Suggested Slow GUI performance for some pages as the lsnodebootdrive command generates unexpected output (show details)
Symptom Performance
Environment SVC systems with CG8 model nodes or older that are running v7.3 or later
Trigger Select Monitor -> Systems page
Workaround Remove and re-add non-DH8 nodes
7.8.1.0 Graphical User Interface
HU01404 All Suggested A node warmstart may occur when a new volume is created using fast format and foreground I/O is submitted to the volume (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger Volume is actively fast formatting when write I/O arrives
Workaround Do not use fast format, or if fast format is used do not submit write I/O to the volume until the format process is complete
7.8.1.0
HU01445 SVC, V7000 Suggested Systems with heavily used RAID-1 or RAID-10 arrays may experience a node warmstart (show details)
Symptom Single Node Warmstart
Environment SVC & V7000 systems running v7.7 or later
Trigger None
Workaround None
7.8.1.0
HU01463 All Suggested SSH Forwarding is enabled on the SSH server (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
7.8.1.0
HU01466 SVC, V7000, V5000 Suggested Stretched cluster and HyperSwap I/O routing does not work properly due to incorrect ALUA data (show details)
Symptom None
Environment Systems running v7.8 or later using HyperSwap
Trigger Move a host to an alternate I/O group
Workaround Use chhost -site to set the correct site
7.8.1.0 Hosts, HyperSwap
HU01470 All Suggested T3 might fail during svcconfig recover -execute while running chemail if the email_machine_address contains a comma (show details)
Symptom None
Environment All systems
Trigger email_machine_address contains a comma with a preceding whitespace
Workaround Modify the config backup or svc.config.recover.sh script by removing trailing whitespace
7.8.1.0 Reliability Availability Serviceability
HU01473 All Suggested EasyTier migrates an excessive number of cold extents to an overloaded nearline array (show details)
Symptom Performance
Environment Systems running v7.6.1 or later using EasyTier
Trigger None
Workaround None
7.8.1.0 EasyTier
HU01487 All Suggested Small increase in read response time for source volumes with additional FlashCopy maps (show details)
Symptom Performance
Environment Systems running v7.5 or later using Global Mirror with Change Volumes
Trigger Add FlashCopy mappings to GMCV source volumes
Workaround Avoid using incremental FlashCopy
7.8.1.0 FlashCopy, Global Mirror with Change Volumes
HU01497 All Suggested A drive can still be offline even though the error is showing as corrected in the Event Log (show details)
Symptom None
Environment Systems running v7.6 or later using DRAID
Trigger None
Workaround Reseat/replace drive
7.8.1.0 Distributed RAID
HU01498 All Suggested GUI may be exposed to CVE-2017-5638 (see Section 3.1) 7.8.1.0 Security
IT19232 V7000, V5000, V3700, V3500 Suggested Storwize systems can report unexpected drive location errors as a result of a RAID issue (show details)
Symptom None
Environment Storwize systems running v7.3 or later
Trigger None
Workaround Replace the drive
7.8.1.0
HU01225 & HU01330 & HU01412 All Critical Node warmstarts due to inconsistencies arising from the way cache interacts with compression (show details)
Symptom Loss of Access to Data
Environment Systems running v7.3 or later with compressed volumes
Trigger None
Workaround None
7.8.0.2 Cache, Compression
HU01410 SVC Critical An issue in the handling of FlashCopy map preparation can cause both nodes in an I/O group to be put into service state (show details)
Symptom Loss of Access to Data
Environment SVC systems using FlashCopy
Trigger None
Workaround None
7.8.0.2 FlashCopy
HU01442 All Critical Upgrading to v7.7.1.5 or v7.8.0.1 with encryption enable will result in multiple Tier 2 recoveries and a loss of access (show details)
Symptom Loss of Access to Data
Environment Systems running v7.5 or earlier
Trigger Upgrade to v7.8.0.1
Workaround None
7.8.0.2 Encryption, System Update
HU00762 All High Importance Due to an issue in the cache component, nodes within an I/O group are not able to form a caching-pair and are serving I/O through a single node (show details)
Symptom Loss of Redundancy
Environment Systems running v7.3 or later
Trigger None
Workaround None
7.8.0.2 Cache
HU01409 All High Importance Cisco Nexus 3000 switches at v5.0(3) have a defect which prevents a config node IP address changing in the event of a fail over (show details)
Symptom Loss of Redundancy
Environment Systems connected to Cisco Nexus 3000 switches
Trigger Config node fail over
Workaround None
7.8.0.2 Reliability Availability Serviceability
HU01426 All High Importance Systems running v7.6.1 or earlier, with compressed volumes, that upgrade to v7.8.0 or later will fail when the first node warmstarts and enters a service state (show details)
Symptom Loss of Redundancy
Environment Systems running v7.6.1 or earlier using compressed volumes
Trigger Upgrade from v7.6.1 or earlier to 7.8.0.0 or later with compressed volumes
Workaround None
7.8.0.2 System Update
HU01432 All Suggested Node warmstart due to an accounting issue within the cache component (show details)
Symptom Single Node Warmstart
Environment Systems running v7.3 or later
Trigger None
Workaround None
7.8.0.2 Cache
HU01459 V7000, V5000 Suggested The event log indicates incorrect enclosure type (show details)
Symptom None
Environment Storwize V5000 and V7000 Gen 2 systems
Trigger None
Workaround Replace the enclosure with a recently manufactured FRU part
7.8.0.2 System Monitoring
IT18752 All Suggested When the config node processes an lsdependentvdisks command, issued via the GUI, that has a large number of objects in its parameters, it may warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.7 or later
Trigger Issue lsdependentvdisks command, via the GUI, for a large number of objects
Workaround Issue those commands, via the CLI
7.8.0.2 Graphical User Interface
HU01382 All HIPER Mishandling of extent migration following a rmarray command can lead to multiple simultaneous node warmstarts with a loss of access (show details)
Symptom Loss of Access to Data
Environment Systems running v7.7.1 or later using DRAID
Trigger Executing a rmarray command
Workaround Avoid using rmarray command
7.8.0.1 Distributed RAID
HU01415 V3700 Critical When a V3700 with 1GE adapters is upgraded to v7.8.0.0 iSCSI hosts will lose access to volumes (show details)
Symptom Loss of Access to Data
Environment V3700 systems running v7.8.0.0 with 1GE adapters
Trigger Upgrade to v7.8.0.0
Workaround None
7.8.0.1 iSCSI, Hosts
HU01193 All HIPER A drive failure whilst an array rebuild is in progress can lead to both nodes in an I/O group warmstarting (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID
Trigger Drive failure
Workaround None
7.8.0.0 Distributed RAID
HU01379 All HIPER Resource leak in the handling of Read Intensive drives leads to offline volumes (show details)
Symptom Offline Volumes
Environment Systems running v7.7.1 or later using read intensive drives
Trigger None
Workaround None
7.8.0.0 Drives
HU00906 All Critical When a compressed volume mirror copy is taken offline, write response times to the primary copy may reach prohibitively high levels leading to a loss of access to that volume (show details)
Symptom Loss of Access to Data
Environment Systems running v7.3 or later that are using volume mirroring with compression
Trigger None
Workaround Only operate mirrored copies on non-compressed volumes
7.8.0.0 Compression, Volume Mirroring
HU01021 & HU01157 All Critical A fault in a backend controller can cause excessive path state changes leading to node warmstarts and offline volumes (show details)
Symptom Offline Volumes
Environment Systems with backend storage
Trigger Faulty backend controller adapters
Workaround None
7.8.0.0 Backend Storage
HU01267 All Critical An unusual interaction between Remote Copy and FlashCopy can lead to both nodes in an I/O group warmstarting (show details)
Symptom Loss of Access to Data
Environment Systems using GMCV
Trigger There may be a number of different scenarios that trigger this issue. Here is one: With GMCV running; Stop GMCV with -access (i.e. svctask stoprcconsistgrp -access); Secondary site becomes primary site and vice versa; Start a FlashCopy on the now primary site; Stop FlashCopy restore before it completes (with -force); Current primary becomes exposed to the issue; Reinstate GMCV (primary becomes secondary again); Stop GMCV with -access again; Both nodes warmstart.
Workaround None
7.8.0.0 Global Mirror with Change Volumes
HU01320 All Critical A rare timing condition can cause hung I/O leading to warmstarts on both nodes in an I/O group. Probability can be increased in the presence of failing drives. (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround Replace faulty drives promptly
7.8.0.0 Hosts
HU01340 All Critical A port translation issue between v7.5 or earlier and v7.7.0 or later requires a Tier 2 recovery to complete an upgrade (show details)
Symptom Loss of Access to Data
Environment Systems running v7.5 or earlier
Trigger Upgrade to v7.7.x
Workaround None
7.8.0.0 System Update
HU01392 All Critical Under certain rare conditions FC mappings not in a consistency group can be added to a special internal consistency group resulting in a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems running v7.7.1 or later
Trigger None
Workaround None
7.8.0.0 FlashCopy
HU01455 All Critical VMWare hosts with ATS enabled can see LUN disconnects to volumes when GMCV is used (show details)
Symptom Loss of Access to Data
Environment Systems running v7.7 or later presenting volumes to VMware host using Global Mirror with Change Volumes
Trigger Enable ATS
Workaround Disable ATS when using Global Mirror with Change Volumes
7.8.0.0 Global Mirror with Change Volumes
HU01519 V7000 Critical One PSU may silently fail leading to the possibility of a dual node reboot (show details)
Symptom Loss of Access to Data
Environment Storwize V7000 Gen 1 systems
Trigger None
Workaround None
7.8.0.0 Reliability Availability Serviceability
HU01635 All Critical A slow memory leak in the host layer can lead to an out-of-memory condition resulting in offline volumes or performance degradation (show details)
Symptom Offline Volumes
Environment Systems running v7.6 or later
Trigger None
Workaround None
7.8.0.0 Hosts, Performance
HU01783 All Critical Replacing a failed drive in a DRAID array, with a smaller drive, may result in multiple Tier 2 recoveries putting all nodes in service state with error 564 and/or 550 (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID
Trigger Replacing a drive with one of less capacity
Workaround Ensure replacement drives are the same capacity
7.8.0.0 Distributed RAID
HU01831 All Critical Cluster-wide warmstarts may occur when the SAN delivers a FDISC frame with an invalid WWPN (show details)
Symptom Loss of Access to Data
Environment Systems running v7.6 or later
Trigger FDISC frame with an invalid WWPN arrives from SAN
Workaround None
7.8.0.0 Reliability Availability Serviceability
HU01177 All High Importance A small timing window issue exists where a node warmstart or power failure can lead to repeated warmstarts of that node until a node rescue is performed (show details)
Symptom Multiple Node Warmstarts
Environment All systems
Trigger None
Workaround None
7.8.0.0 Reliability Availability Serviceability
HU01223 All High Importance The handling of a rebooted nodes return to the cluster can occasionally become delayed resulting in a stoppage of inter-cluster relationships (show details)
Symptom Loss of Redundancy
Environment Systems running v7.3 or later using MetroMirror
Trigger None
Workaround None
7.8.0.0 Metro Mirror
HU01254 SVC High Importance A fluctuation of input AC power can cause a 584 error on a node (show details)
Symptom Loss of Redundancy
Environment SVC CF8/CG8 systems
Trigger None
Workaround None
7.8.0.0 Reliability Availability Serviceability
HU01268 V7000, V5000, V3700, V3500 High Importance Upgrade to 7.7.x fails on Storwize systems in the replication layer where a T3 recovery was performed in the past (show details)
Symptom None
Environment Storwize systems in the replication layer that have experienced a T3 recovery in the past
Trigger Upgrade to v7.7.x
Workaround None
7.8.0.0 System Update
HU01347 All High Importance During an upgrade to v7.7.1 a deadlock in node communications can occur leading to a timeout and node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v7.7.0 or earlier using SE volumes
Trigger Upgrade to v7.7.1
Workaround None
7.8.0.0 Thin Provisioning
HU01381 All High Importance A rare timing issue in FlashCopy may lead to a node warmstarting repeatedly and then entering a service state (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v7.2 or later using FlashCopy
Trigger None
Workaround None
7.8.0.0 FlashCopy
HU01402 V7000 High Importance Nodes can power down unexpectedly as they are unable to determine from their partner whether power is available (show details)
Symptom Loss of Redundancy
Environment V7000 Gen 1 systems running v7.6 or later
Trigger None
Workaround None
7.8.0.0 Reliability Availability Serviceability
HU01488 V7000, V5000 High Importance SAS transport errors on an enclosure slot have the potential to affect an adjacent slot leading to double drive failures (show details)
Symptom Loss of Redundancy
Environment Storwize V5000 and V7000 systems running v7.7 or later
Trigger None
Workaround None
7.8.0.0 Drives
IT14917 All High Importance Node warmstarts due to a timing window in the cache component. For more details refer to this Flash (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v7.4 or later
Trigger None
Workaround None
7.8.0.0 Cache
IT16012 SVC High Importance Internal node boot drive RAID scrub process at 1am every Sunday can impact system performance (show details)
Symptom Performance
Environment Systems running v7.3 or later
Trigger Internal node boot drive RAID scrub process
Workaround Try to avoid performing high I/O workloads (including copy services) at 1am on Sundays
7.8.0.0 Performance
IT17564 All High Importance All nodes in an I/O group may warmstart when a DRAID array experiences drive failures (show details)
Symptom Loss of Access to Data
Environment Systems with DRAID arrays
Trigger Failing drives
Workaround None
7.8.0.0 Distributed RAID
HU00831 All Suggested Single node warmstart due to hung I/O caused by cache deadlock (show details)
Symptom Single Node Warmstart
Environment Systems running v7.2 or later
Trigger Hosts sending many large block IOs that use many credits per I/O
Workaround None
7.8.0.0 Cache
HU01098 All Suggested Some older backend controller code levels do not support C2 commands resulting in 1370 entries in the Event Log for every detectmdisk (show details)
Symptom None
Environment Systems running v7.6 or later attached to backend controllers running older code levels
Trigger Issue CLI command detectmdisk
Workaround None
7.8.0.0 Backend Storage
HU01213 All Suggested The LDAP password is visible in the auditlog (show details)
Symptom None
Environment Systems running v7.7 or later
Trigger Setup LDAP authentication
Workaround None
7.8.0.0 LDAP
HU01228 All Suggested Automatic T3 recovery may fail due to the handling of quorum registration generating duplicate entries (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
7.8.0.0 Reliability Availability Serviceability
HU01229 V7000, V5000, V3700, V3500 Suggested The DMP for a 3105 event does not identify the correct problem canister (show details)
Symptom None
Environment Storwize systems
Trigger None
Workaround Problem canister identified in Event log
7.8.0.0 GUI Fix Procedure
HU01230 All Suggested A host aborting an outstanding logout command can lead to a single node warmstart (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
7.8.0.0
HU01247 All Suggested When a FlashCopy consistency group is stopped more than once in rapid succession a node warmstart may result (show details)
Symptom Single Node Warmstart
Environment Systems using FlashCopy
Trigger Stop the same FlashCopy consistency group twice in rapid succession using -force on second attempt
Workaround Avoid stopping the same flash copy consistency group more than once
7.8.0.0 FlashCopy
HU01264 All Suggested Node warmstart due to an issue in the compression optimisation process (show details)
Symptom Single Node Warmstart
Environment Systems that are using compressed volumes
Trigger None
Workaround None
7.8.0.0 Compression
HU01269 All Suggested A rare timing conflict between two process may lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
7.8.0.0
HU01304 All Suggested SSH authentication fails if multiple SSH keys are configured on the client (show details)
Symptom None
Environment Systems running v7.7 or later
Trigger On client create 2 x SSH key pairs (neither of which are configured on the SVC/Storwize) then connect to SVC/Storwize system
Workaround Remove other keys from the client or explicitly use password authentication on the client: ssh -o PreferredAuthentications=password <user>@<host>
7.8.0.0
HU01323 All Suggested Systems using Volume Mirroring that upgrade to v7.7.1.x and have a storage pool go offline may experience a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.7.1 or later using volume mirroring
Trigger 1. Mirrored volumes are created at a pre-7.7.1 code level; 2. The system is upgraded to 7.7.1.1 or later; 3. A storage pool goes offline before the nodes next restart.
Workaround None
7.8.0.0 Volume Mirroring
HU01370 All Suggested lsfabric command may not list all logins when it is used with parameters (show details)
Symptom None
Environment Systems running v7.7 or later
Trigger Enter lsfabric with parameters
Workaround Enter lsfabric without parameters and use client tools such as grep for filtering
7.8.0.0 Command Line Interface
HU01374 All Suggested Where an issue with Global Mirror causes excessive I/O delay, a timeout may not function resulting in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.7.0 or later using Global Mirror
Trigger None
Workaround None
7.8.0.0 Global Mirror
HU01399 All Suggested For certain config nodes the CLI Help commands may not work (show details)
Symptom None
Environment All systems
Trigger None
Workaround Use Knowledge Center
7.8.0.0 Command Line Interface
IT17302 V5000, V3700, V3500 Suggested Unexpected 45034 1042 entries in the Event Log (show details)
Symptom None
Environment Systems running v7.7 or later
Trigger None
Workaround None
7.8.0.0 System Monitoring
IT18086 All Suggested When a volume is moved between I/O groups a node may warmstart (show details)
Symptom None
Environment Systems running v7.7.1 or later
Trigger No warmstarts since upgrade. Move a volume between I/O groups
Workaround None
7.8.0.0

4. Useful Links

Description Link
Support Websites
Update Matrices, including detailed build version
Support Information pages providing links to the following information:
  • Interoperability information
  • Product documentation
  • Limitations and restrictions, including maximum configuration limits
Supported Drive Types and Firmware Levels
SAN Volume Controller and Storwize Family Inter-cluster Metro Mirror and Global Mirror Compatibility Cross Reference
Software Upgrade Test Utility
Software Upgrade Planning