Release Note for V9000 Family Block Storage Products


This release note applies to the following systems: This is the release note for the 8.2.1 release and details the issues resolved in all Program Temporary Fixes (PTFs) between 8.2.1.0 and 8.2.1.18. This document will be updated with additional information whenever a PTF is released.

This document was last updated on 7th February 2023.

  1. New Features
  2. Known Issues and Restrictions
  3. Issues Resolved
    1. Security Issues Resolved
    2. APARs Resolved
  4. Supported upgrade paths
  5. Useful Links

1. New Features

The following new features have been introduced in the 8.2.1 release: The following feature has been introduced in the 8.2.1.3 release:

2. Known Issues and Restrictions

Details Introduced

Customers with more than 5 x non-NVMe over FC hosts (i.e FC SCSI or iSCSI) in an I/O group must not attach any NVMe over FC hosts to that I/O group.
Customers with more than 20 x non-NVMe over FC hosts (i.e FC SCSI or iSCSI) in a cluster must not attach any NVMe over FC hosts to that cluster.

These limits will not be policed by the Spectrum Virtualize software. Any configurations that exceed these limits will experience significant adverse performance impact.

These limits will be lifted in a future major release.

8.2.1.0

Customers using Transparent Cloud Tiering should not upgrade to v8.2.1.0.

This is a restriction that may be lifted in a future PTF.

8.2.1.0

Customers using iSCSI to virtualize backend controllers should not upgrade to v8.2.0 or later

This is a restriction that may be lifted in a future PTF.

8.2.0.0

Customers using the REST API to list more than 2000 objects may experience a loss of service from the API as it restarts due to memory constraints.

This is a restriction that may be lifted in a future PTF.

8.1.3.0

It is not possible to access the REST API using a cluster's IPv6 address.

This is a restriction that may be lifted in a future PTF.

8.1.3.0

Customers upgrading systems with more than 64GB of RAM to v8.1 or later will need to run chnodehw to enable access to the extra memory above 64GB.

Under some circumstances it may also be necessary to remove and re-add each node in turn.

8.1.0.0

Validation in the Upload Support Package feature will reject the new case number format in the PMR field.

This is a known issue that may be lifted in a future PTF. The fix can be tracked using APAR HU02392.

7.8.1.0

Systems, with NPIV enabled, presenting storage to SUSE Linux Enterprise Server (SLES) or Red Hat Enterprise Linux (RHEL) hosts running the ibmvfc driver on IBM Power can experience path loss or read-only file system events.

This is cause by issues within the ibmvfc driver and VIOS code.

Refer to this troubleshooting page for more information.

n/a
If an update stalls or fails then contact IBM Support for further assistance n/a
The following restrictions were valid but have now been lifted

Customers with direct attached external storage controllers cannot upgrade to v8.2.1.6. This is most likely to impact customers with 2 x1 Direct Attach V9000 systems.

This has been resolved, under APAR HU02077, in v8.2.1.8.

Please note that v8.2.1.5, or earlier, is not exposed to this restriction.

8.2.1.6

Customers using systems with attached AE2 enclosures cannot upgrade to versions 8.2.1.0 - 8.2.1.4, but rather, must upgrade directly to v8.2.1.5 or later.

8.2.1.0

With Gemalto SafeNet KeySecure, the chkeyserverkeysecure -username <username> command is used to set the KeySecure username credential. If this is changed to a username that is not recognised by the key server to be the valid username, associated with the Spectrum Virtualize encryption key, then a subsequent re-key operation can cause key servers to appear offline.

This is a an issue that will be resolved in a future PTF.

8.2.1.0

3. Issues Resolved

This release contains all of the fixes included in the 8.1.3.1 release, plus the following additional fixes.

A release may contain fixes for security issues, fixes for APARs or both. Consult both tables below to understand the complete set of fixes included in the release.

3.1 Security Issues Resolved

Security issues are documented using a reference number provided by "Common Vulnerabilities and Exposures" (CVE).
CVE Identifier Link for additional Information Resolved in
CVE-2023-2597 7065011 8.2.1.18
CVE-2022-21626 6858041 8.2.1.17
CVE-2022-43873 6858047 8.2.1.17
CVE-2022-0778 6622017 8.2.1.16
CVE-2021-35603 6622019 8.2.1.16
CVE-2021-35550 6622019 8.2.1.16
CVE-2018-25032 6622021 8.2.1.16
CVE-2021-38969 6584337 8.2.1.15
CVE-2021-29873 6497111 8.2.1.14
CVE-2020-2781 6445063 8.2.1.12
CVE-2020-13935 6445063 8.2.1.12
CVE-2020-14577 6445063 8.2.1.12
CVE-2020-14578 6445063 8.2.1.12
CVE-2020-14579 6445063 8.2.1.12
CVE-2019-5544 6250889 8.2.1.11
CVE-2019-2964 6250887 8.2.1.11
CVE-2019-2989 6250887 8.2.1.11
CVE-2018-12404 6250885 8.2.1.11
CVE-2019-11477 1164286 8.2.1.6
CVE-2019-11478 1164286 8.2.1.6
CVE-2019-11479 1164286 8.2.1.6
CVE-2019-2602 1073958 8.2.1.6
CVE-2018-3180 ibm10884526 8.2.1.4
CVE-2018-12547 ibm10884526 8.2.1.4
CVE-2008-5161 ibm10874368 8.2.1.2
CVE-2018-5391 ibm10872368 8.2.1.2
CVE-2018-11776 ibm10741137 8.2.1.0
CVE-2017-17449 ibm10872364 8.2.1.0
CVE-2017-18017 ibm10872364 8.2.1.0
CVE-2017-17833 ibm10872546 8.2.1.0
CVE-2018-11784 ibm10872550 8.2.1.0
CVE-2018-5732 ibm10741135 8.2.1.0
CVE-2018-1517 ibm10872456 8.2.1.0
CVE-2018-2783 ibm10872456 8.2.1.0
CVE-2018-12539 ibm10872456 8.2.1.0
CVE-2018-1775 ibm10872486 8.2.1.0

3.1 APARs and Flashes Resolved

For more details refer to the following Flash 
Reference Severity Description Resolved in Feature Tags
HU02327 S1 Using addvdiskcopy in conjunction with expandvdisk with format may result in the original being overwritten, by the new copy, producing blank copies.  For more details refer to the following Flash  8.2.1.15 Volume Mirroring
HU02400 S1 A problem in the virtualization component of the system can cause a migration IO to be submitted in an incorrect context resulting in a node warmstart.  In some cases it is possible that this IO has been submitted to an incorrect location on the backend,  which can cause data corruption of an isolated small area 8.2.1.15 Storage Virtualisation
HU02342 S1 Occasionally when an offline drive returns to online state later than its peers in the same RAID array there can be multiple node warmstarts that send nodes into a service state 8.2.1.15 RAID
HU02374 S1 Hosts with Emulex 16Gbps HBAs may become unable to communicate with a system with 8Gbps Fibre Channel ports, after the host HBA is upgraded to firmware version 12.8.364.11. This does not apply to systems with 16Gb or 32Gb Fibre Channel ports 8.2.1.15 Hosts
HU02393 S1 Automatic resize of compressed/thin volumes may fail causing warmstarts on both nodes in an I/O group 8.2.1.15 Storage Virtualisation
HU02401 S1 EasyTier can move extents between identical mdisks until one runs out of space 8.2.1.15 EasyTier
HU02406 S1 An interoperability issue between Cisco NX-OS firmware and the Spectrum Virtualize Fibre Channel driver can cause a node warmstart on NPIV failback (for example during an upgrade) with the potential for a loss of access.  For more details refer to the following Flash  8.2.1.15 Interoperability
IT38015 S2 During RAID rebuild or copyback on systems with 16gb or less of memory, cache handling can lead to a deadlock which results in timeouts 8.2.1.15 RAID
HU02366 S3 Slow internal resource reclamation by the RAID component can cause a node warmstart 8.2.1.15 RAID
HU02433 S3 When a BIOS upgrade occurs excessive tracefile entries can be generated 8.2.1.15 System Update
HU02527 S3 If more than 15 SAS enclosures are attached, callhome may fail due to the email being too large for the server to process. 8.2.1.15 System Monitoring
HU02277 S1 HIPER (Highly Pervasive): RAID parity scrubbing can become stalled causing an accumulation of media errors leading to multiple drive failures with the possibility of data integrity loss.  For more details refer to the following  Flash  8.2.1.12 RAID
HU01968 & HU02215 S1 An upgrade may fail due to corrupt hardened data in a node. This can affect an I/O group 8.2.1.12 System Update
HU02058 S1 Changing a remote copy relationship from GMCV to MM or GM can result in a Tier 2 recovery 8.2.1.12 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU02213 S1 A Hot Spare Node (HSN) timing window issue can, during an HSN activation or deactivation,  cause the cluster to broadcast an invalid VPD update to other clusters on the SAN. This may trigger a Tier 2 recovery on the other cluster.  For more details refer to the following Flash  8.2.1.12 Hot Spare Node
HU02266 S1 An issue in auto-expand can cause expansion to fail and the volume to be taken offline 8.2.1.12 Thin Provisioning
HU02295 S1 When upgrading from v8.2.1 or v8.3, in the presence of hot spare nodes, an issue with the handling of node metadata may cause a Tier 2 recovery 8.2.1.12 System Update
HU02429 S1 System can go offline shortly after changing the SMTP settings using the chemailserver command via the GUI 8.2.1.12 System Monitoring
HU02132 S2 Removing a thin-provisioned volume and then immediately creating one of the same size may cause node warmstarts 8.2.1.12 Thin Provisioning
HU02156 S2 Global Mirror environments may experience more frequent 1920 events due to writedone message queuing 8.2.1.12 Global Mirror
HU02164 S2 An issue in Remote Copy may cause a loss of hardened data when a node is warmstarted 8.2.1.12 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU02176 S2 During upgrade a node may limit the number of target ports it reports causing a failover contradiction on hosts 8.2.1.12 Hosts
HU02184 S1 When a 3PAR controller experiences a fault that prevents normal I/O processing it may issue a SCSI TARGET RESET command. This command is not supported and may cause multiple node asserts, possibly cluster-wide 8.2.1.12 Backend Storage
HU02200 S2 When upgrading from v8.1 or earlier to v8.2.1 or later a remote copy issue may cause a node warmstart, stalling the upgrade 8.2.1.12 System Update
HU02201 & HU02221 S2 Shortly after upgrading drive firmware, specific drive models can fail due to Too many long IOs to drive for too long errors 8.2.1.12 Drives
HU02227 S2 Certain I/O patterns can cause compression hardware to post errors. When those errors exceed a threshold the node can be taken offline 8.2.1.12 Compression
HU02288 S2 A node might fail to come online after a reboot or warmstart such as during an upgrade 8.2.1.12 Reliability Availability Serviceability
HU02048 S3 An issue in the handling of ATS commands from VMware hosts can cause a single node warmstart 8.2.1.12 Hosts
HU02142 S3 It is possible for a backend unmap process to become stalled, preventing system configuration changes from completing 8.2.1.12 Distributed RAID
HU02157 S3 Issuing a mkdistributedarray command may result in a node warmstart 8.2.1.12 Distributed RAID
HU02241 S3 IP Replication can fail to create IP partnerships via the secondary cluster management IP 8.2.1.12 IP Replication
HU02244 S3 False positive node error 766 (depleted CMOS battery) messages may appear in the Event Log 8.2.1.12 System Monitoring
HU02292 & HU02308 S3 The use of maximum replication delay within Global Mirror may occasionally cause a node warmstart 8.2.1.12 Global Mirror
HU02332 & HU02336 S3 When an I/O is received, from a host, with invalid or inconsistent SCSI data but a good checksum it may cause a node warmstart 8.2.1.12 Hosts
HU02354 S3 An issue in the handling of read transfers may cause hung host IOs leading to a node warmstart 8.2.1.12 Hosts
HU02358 S3 An issue in Remote Copy, that stalls a switch of direction, can cause I/O timeouts leading to a node warmstart 8.2.1.12 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01894 S1 HIPER (Highly Pervasive): After node reboot, or warmstart, some volumes accessed by AIX, VIO or VMware hosts, may experience stuck SCSI2 reservations on the NPIV failover ports of the partner node. This can cause a loss of access to data 8.2.1.11 Hosts
HU02141 S1 HIPER (Highly Pervasive): An issue in the max replication delay function may trigger a Tier 2 recovery, after posting multiple 1920 errors in the Event Log.  For more details refer to the following Flash  8.2.1.11 Global Mirror
HU02205 S1 HIPER (Highly Pervasive): Incremental FlashCopy targets can be corrupted when the FlashCopy source is a target of a remote copy relationship 8.2.1.11 FlashCopy, Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU02212 S1 HIPER (Highly Pervasive): Remote Copy secondary may have inconsistent data following a stop with -access due to a missing bitmap merge from FlashCopy to Remote Copy For more details refer to the following Flash  8.2.1.11 Global Mirror with Change Volumes, HyperSwap
HU02237 S1 HIPER (Highly Pervasive): Under a rare and complicated set of conditions, a RAID 1 or RAID 10 array may drop a write, causing undetected data corruption.  For more details refer to the following Flash  8.2.1.11 RAID
HU02238 S1 HIPER (Highly Pervasive): Force-stopping a FlashCopy map, where the source volume is a Metro or Global Mirror target volume, may cause other FlashCopy maps to return invalid data if they are not 100% copied, in specific configurations.  For more details refer to the following Flash 8.2.1.11 RAID
HU01921 S1 Where FlashCopy mapping targets are also in remote copy relationships there may be node warmstarts with a temporary loss of access to data 8.2.1.11 FlashCopy, Global Mirror, Metro Mirror
HU01924 S1 Migrating extents to an MDisk, that is not a member of an MDisk group, may result in a Tier 2 recovery 8.2.1.11 Thin Provisioning
HU01970 S1 When a GMCV relationship is stopped with the -access option, and the secondary volume is immediately deleted with -force, then all nodes may repeatedly warmstart 8.2.1.11 Global Mirror with Change Volumes
HU02005 S1 An issue in the background copy process prevents grains, above a 128TB limit, from being cleaned properly. As a consequence there may be multiple node warmstarts with the potential for a loss of access to data 8.2.1.11 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU02054 S1 The event log handler maintains a second list of events. On rare occasions, for log full events, these lists can get out of step, resulting in a Tier 2 recovery 8.2.1.11 System Monitoring
HU02065 S1 Mishandling of Data Reduction Pool allocation request rejections can lead to node warmstarts that can take an MDisk group offline 8.2.1.11 Data Reduction Pools
HU02069 S1 When a SCSI command, containing an invalid byte, is received there may be a node warmstart. This can affect both nodes, in an I/O group, at the same time 8.2.1.11 Hosts
HU02097 S1 Workloads, with data that is highly suited to deduplication, can provoke high CPU utilisation, as multiple destinations try to dedupe to one source. This adversely impacts performance with the possibility of offline MDisk groups 8.2.1.11 Data Reduction Pools
HU02106 S1 Multiple node warmstarts, in quick succession, can cause the partner node to lease expire 8.2.1.11 IP Quorum
HU02108 S1 Deleting a managed disk group, with -force, may cause multiple warmstarts with the possibility of a loss of access to data 8.2.1.11 Data Reduction Pools
HU02109 S1 Free extents may not be unmapped after volume deletion, or migration, resulting in out-of-space conditions on backend controllers 8.2.1.11 Backend Storage, SCSI Unmap
HU02135 S1 Removing multiple IQNs for an iSCSI host can result in a Tier 2 recovery 8.2.1.11 iSCSI
HU02138 S1 An issue in Data Reduction Pool garbage collection can cause I/O timeouts leading to an offline pool 8.2.1.11 Data Reduction Pools
HU02154 S1 If a node is rebooted, when remote support is enabled, then all other nodes will warmstart 8.2.1.11 Support Remote Assist
HU02155 S1 Upgrading to v8.2.1 may result in offline managed disk groups and OOS events (1685/1687) appearing in the Event Log 8.2.1.11 Data Reduction Pools
HU02197 S1 Bulk volume removals can adversely impact related FlashCopy mappings leading to a Tier 2 recovery 8.2.1.11 FlashCopy
HU02222 S1 Where the source volume of an incremental FlashCopy map is also a Metro or Global Mirror target volume that is using a change volume or is a Hyperswap volume, then there is a possibility that not all data will be copied to the FlashCopy target. For more details refer to the following Flash  8.2.1.11 Global Mirror with Change Volumes
IT25367 S1 A T2 recovery may occur when an attempt is made to upgrade, or downgrade, the firmware for an unsupported drive type 8.2.1.11 Drives
IT31113 S1 After a manual power off and on, of a system, both nodes, in an I/O group, may repeatedly assert into a service state 8.2.1.11 RAID
HU01923 S2 An issue in the way Global Mirror handles write sequence numbers >512 may cause multiple node warmstarts 8.2.1.11 Global Mirror
HU02080 S2 When a Data Reduction Pool is running low on free space, the credit allocation algorithm, for garbage collection, can be exposed to a race condition, adversely affecting performance 8.2.1.11 Data Reduction Pools
HU02123 S2 For direct-attached hosts, a race condition between the FLOGI and Link UP processes can result in FC ports not coming online 8.2.1.11 Hosts
HU02149 S2 When an Enhanced Stretch Cluster is using NPIV, in transitional mode, the path priority is not being reported correctly to some hosts 8.2.1.11 Hosts
HU01868 S3 After deleting an encrypted external MDisk, it is possible for the 'encrypted' status of volumes to change to 'no', even though all remaining MDisks are encrypted 8.2.1.11 Encryption
HU01917 S3 Chrome browser support requires a self-signed certificate to include subject alternate name 8.2.1.11 Graphical User Interface
HU01931 S3 Where a high rate of CLI commands are received, it is possible for inter-node processing code to be delayed which results in a small increase in receive queue time on the config node 8.2.1.11 Performance
HU02111 S3 An issue with how Data Reduction Pools handle data, at the sub-extent level, may result in a node warmstart 8.2.1.11 Data Reduction Pools
HU02124 S3 Due to an issue with FCM thin provisioning calculations the GUI may incorrectly display volume capacity and capacity savings as zero 8.2.1.11 System Monitoring
HU02137 S3 An issue with support for target resets in Nimble Storage controllers may cause a node warmstart 8.2.1.11 Backend Storage
HU02173 S3 During a pending fabric login, when an abort is received, it is possible for a related entry in the WWPN table to not be removed. The node will warmstart to clear this condition. 8.2.1.11 Reliability Availability Serviceability
HU02183 S3 An issue in the way inter-node communication is handled can lead to a node warmstart 8.2.1.11 Reliability Availability Serviceability
HU02190 S3 Error 1046 not triggering a Call Home even though it is a hardware fault 8.2.1.11 System Monitoring
HU02214 S3 Under a certain I/O pattern it is possible for metadata management in Data Reduction Pools to become inconsistent leading to a node warmstart 8.2.1.11 Data Reduction Pools
HU02143 S2 The performance profile, for some enterprise tier drives, may not correctly match the drives capabilities leading to that tier being overdriven 8.2.1.10 EasyTier
HU02104 S1 HIPER (Highly Pervasive): An issue in the RAID component, in the presence of very high I/O workload and the exhaustion of cache resources, can see a deadlock condition occurring which prevents further I/O processing. The system detects this issue and takes the storage pool offline for a six minute period, to clear the problem. The pool is then brought online automatically, and normal operation resumes. For more details refer to the following Flash  8.2.1.9 RAID, Distributed RAID
HU02102 S3 Excessive processing time required for FlashCopy bitmap operations, associated with large (>20TB) Global Mirror change volumes, may lead to a node warmstart 8.2.1.9 Global Mirror with Change Volumes
HU02131 S3 When changing DRAID configuration, for an array with an active workload, a deadlock condition can occur resulting in a single node warmstart 8.2.1.9 Distributed RAID
HU02126 S3 There is a low probability that excessive SSH connections may trigger a single node warmstart on the configuration node 8.2.1.9 Command Line Interface
HU02064 S1 HIPER (Highly Pervasive): An issue in the firmware for compression accelerator cards can cause offline compressed volumes. For more details refer to the following Flash 8.2.1.8 Compression
HU02083 S1 HIPER (Highly Pervasive): During DRAID rebuilds, an issue in the handling of memory buffers can lead to multiple node warmstarts and a loss of access to data For more details refer to the following Flash 8.2.1.8 Distributed RAID
HU01967 S1 When I/O, in remote copy relationships, experiences delays (1720 and/or 1920 errors are logged) an I/O group may warmstart 8.2.1.8 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU02036 S1 It is possible for commands, that alter pool-level extent reservations (i.e. migratevdisk or rmmdisk), to conflict with an ongoing EasyTier migration, resulting in a Tier 2 recovery 8.2.1.8 EasyTier
HU02044 S1 Multiple DRAID arrays can, where one is performing a rebuild, be exposed to a RAID deadlock condition resulting in multiple node warmstarts and a loss of access to data 8.2.1.8 Distributed RAID
HU02063 S1 HyperSwap clusters with only two surviving nodes may experience warmstarts on both of those nodes where rcbuffersize is set to 512MB 8.2.1.8 HyperSwap
HU02077 S1 Upgrading to 8.2.1.6 or 8.3.0.0 can cause a loss of access to direct-attached Fibre Channel controllers 8.2.1.8 Backend Storage
HU02086 S1 An issue, in IP Quorum, may cause a Tier 2 recovery, during initial connection to a candidate device 8.2.1.8 IP Quorum
HU02089 S1 Due to changes to quorum management, during an upgrade to v8.2.x, or later, there may be multiple warmstarts, with the possibility of a loss of access to data 8.2.1.8 System Update
HU02121 S1 When the system changes from copyback to rebuild a failure to clear related metadata can cause multiple node warmstarts, with the possibility of a loss of access 8.2.1.8 Distributed RAID
IT26257 S1 Starting a relationship, when the remote volume is offline, may result in a T2 recovery 8.2.1.8 HyperSwap
IT30595 S1 A resource shortage in the RAID component can cause MDisks to be taken offline 8.2.1.8 RAID
FLASH-29324 S2 Reading enclosure VPD can cause timeouts leading to node warmstarts 8.2.1.8 Reliability Availability Serviceability
HU01836 S2 When an auxiliary volume is moved an issue with pausing the master volume can lead to node warmstarts 8.2.1.8 HyperSwap
HU02049 S2 GUI session handling has an issue that can generate many exceptions, adversely impacting GUI performance 8.2.1.8 Graphical User Interface
HU01880 S3 When a write, to a secondary volume, becomes stalled, a node at the primary site may warmstart 8.2.1.8 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01936 S3 When shrinking a volume, that has host mappings, there may be recurring node warmstarts 8.2.1.8 Cache
HU02021 S3 Disabling garbage collection may cause a node warmstart 8.2.1.8 Data Reduction Pools
HU02085 S3 Freeze time of Global Mirror remote copy consistency groups may not be updated correctly in certain scenarios 8.2.1.8 Global Mirror
IT30448 S3 If an IP Quorum app is killed, during the commit phase of a code upgrade, then that offline IP Quorum device cannot be removed, post upgrade 8.2.1.8 IP Quorum
HU01998 S1 HIPER (Highly Pervasive): All SCSI command types can set volumes as busy resulting in I/O timeouts, and multiple node warmstarts, with the possibility of an offline I/O group. For more details refer to the following Flash  8.2.1.6 Hosts
HU02014 S1 HIPER (Highly Pervasive): After a loss of power, where an AC3 node has a dead CMOS battery, it will fail to restart correctly. It is possible for both nodes in an I/O group to experience this issue 8.2.1.6 Reliability Availability Serviceability
HU01888 & HU01997 S1 An issue with restore mappings, in the FlashCopy component, can cause an I/O group to warmstart 8.2.1.6 FlashCopy
HU01933 S1 Under rare circumstances the Data Reduction Pool deduplication rehoming process can become truncated. Subsequent detection of inconsistent metadata can lead to offline Data Reduction Pools 8.2.1.6 Data Reduction Pools, Deduplication
HU01985 S1 As a consequence of a Data Reduction Pool recovery bad metadata may be created. When the region of disk associated with the bad metadata is accessed there may be an I/O group warmstarts 8.2.1.6 Data Reduction Pools
HU01989 S1 For large drives, bitmap scanning, during a rebuild, can timeout resulting in multiple node warmstarts, possibly leading to offline I/O groups 8.2.1.6 Distributed RAID
HU02016 S1 A memory leak in the component that handles thin-provisioned MDisks can lead to an adverse performance impact with the possibility of offline MDisks 8.2.1.6 Backend Storage
HU02027 S1 Fabric congestion can cause internal resource constraints, in 16Gb HBAs, leading to lease expiries 8.2.1.6 Reliability Availability Serviceability
HU02043 S1 Collecting a snap can cause nodes to run out of boot drive space and go offline with node error 565 8.2.1.6 Support Data Collection
HU02045 S1 When a node is removed from the cluster, using CLI, it may still be shown as online in the GUI. If an attempt is made to shutdown this node, from the GUI, whilst it appears to be online, then the whole cluster will shutdown 8.2.1.6 Graphical User Interface
HU01890 S2 FlashCopy mappings, from master volume to primary change volume, may become stalled when a T2 recovery occurs whilst the mappings are in a 'copying' state 8.2.1.6 Global Mirror with Change Volumes
HU02037 S2 A FlashCopy consistency group, with a mix of mappings in different states, cannot be stopped 8.2.1.6 FlashCopy
HU02046 S2 A timer issue within the support data collection function can cause truncated system report files for systems with AE2 enclosures 8.2.1.6 Support Data Collection
HU02047 S2 When a system is upgraded to 8.2.1.5, a large number of events can be sent to call home 8.2.1.6 System Monitoring
HU02055 S2 Creating a FlashCopy snapshot, in the GUI, does not set the same preferred node for both source and target volumes. This may adversely impact performance 8.2.1.6 FlashCopy
HU02072 S2 An issue in the handling of email transmission can write a large file to the node boot drive. If this causes the boot drive to become full, the node will go offline with error 565 8.2.1.6 System Monitoring
FLASH-28947 S3 Active rebuild commands stuck after cross-bar failure 8.2.1.6 Reliability Availability Serviceability
FLASH-29059 S3 Keep fans running in a canister that powers down to standby 8.2.1.6 Reliability Availability Serviceability
FLASH-29060 S3 DMP support for keeping fans running in a canister that powers down to standby 8.2.1.6 GUI Fix Procedure
FLASH-29061 S3 On system shutdown, power down fans 8.2.1.6 Reliability Availability Serviceability
FLASH-29062 S3 Ramp fans to maximum speed when canister is missing 8.2.1.6 Reliability Availability Serviceability
HU01843 S3 A node hardware issue can cause a CLI command to timeout resulting in a node warmstart 8.2.1.6 Command Line Interface
HU01892 S3 LUNs of greater than 2TB, presented by HP XP7 storage controllers, are not supported 8.2.1.6 Backend Storage
HU01974 S3 With all Remote Support Assistant connections closed, the GUI may show that a connection is still in progress 8.2.1.6 System Monitoring
HU01978 S3 Unable to create HyperSwap volumes. The mkvolume command fails with CMMVC7050E error 8.2.1.6 HyperSwap
HU01979 S3 The figure for used_virtualization, in the output of a lslicense command, may be unexpectedly large 8.2.1.6 Command Line Interface
HU01982 S3 In an environment, with multiple IP Quorum servers, if the quorum component encounters a duplicate UID then a node may warmstart 8.2.1.6 IP Quorum
HU01983 S3 Improve debug data capture to assist in determining the reason for a Data Reduction Pool to be taken offline 8.2.1.6 Data Reduction Pools
HU01986 S3 An accounting issue in the FlashCopy component may cause node warmstarts 8.2.1.6 FlashCopy
HU01991 S3 An issue in the handling of extent allocation, in the Data Reduction Pool component, can cause a node warmstart 8.2.1.6 Data Reduction Pools
HU02029 S3 An issue with the SSMTP process may result in failed callhome, inventory reporting and user notifications. A testemail command will fail with a CMMVC9051E error 8.2.1.6 System Monitoring
HU02039 S3 An issue in the management steps of Data Reduction Pool recovery may lead to a node warmstart 8.2.1.6 Data Reduction Pools
HU02067 S3 If multiple recipients are specified, for callhome emails, then no callhome emails will be sent 8.2.1.6 System Monitoring
HU02129 S3 GUI drive filtering fails with "An error occurred loading table data" 8.2.1.6 Graphical User Interface
HU02007 S1 HIPER (Highly Pervasive): During volume migration an issue in the handling of old to new extents transfer can lead to cluster-wide warmstarts 8.2.1.5 Storage Virtualisation
HU02009 S1 Systems which are using Data Reduction Pools, with the maximum possible extent size of 8GB, and which experience a very specific I/O workload, may experience an issue due to garbage collection. This can cause repeated node warmstarts and loss of access to data 8.2.1.5 Data Reduction Pools
HU02011 S2 When a node warmstart occurs on a system using Data Reduction Pools, there is a small possibility that the node will not automatically return online. If the partner node is also offline, this can cause temporary loss of access to data 8.2.1.5 Data Reduction Pools
HU02012 S2 Under certain I/O workloads the garbage collection process can adversely impact volume write response times 8.2.1.5 Data Reduction Pools
HU01865 S1 HIPER (Highly Pervasive): When creating a HyperSwap relationship, using addvolumecopy (or similar methods), the system should perform a synchronisation operation to copy the data of the original copy to the new copy. In some cases this synchronisation is skipped, leaving the new copy with bad data (all zeros) 8.2.1.4 HyperSwap
HU01918 S1 HIPER (Highly Pervasive): Where Data Reduction Pools have been created on earlier code levels, upgrading the system, to an affected release, can cause an increase in the level of concurrent flushing to disk. This may result in a loss of access to data. For more details refer to the following Flash  8.2.1.4 Data Reduction Pools
HU02008 S1 HIPER (Highly Pervasive): When a DRAID rebuild occurs, occasionally a RAID deadlock condition can be triggered by a particular type of I/O workload. This can lead to repeated node warmstarts and a loss of access to data 8.2.1.4 Distributed RAID
HU01887 S1 In circumstances where host configuration data becomes inconsistent, across nodes, an issue in the CLI policing code may cause multiple warmstarts 8.2.1.4 Command Line Interface, Host Cluster
HU01900 S1 Executing a command, that can result in a shared mapping being created or destroyed, for an individual host, in a host cluster, without that command applying to all hosts in the host cluster, may lead to multiple node warmstarts with the possibility of a T2 recovery 8.2.1.4 Host Cluster
HU01910 S1 When FlashCopy mappings are created, with a grain size of 64KB, it is possible for an overflow condition in the bitmap to occur. This can resulting in multiple node warmstarts with a possible loss of access to data 8.2.1.4 FlashCopy
HU01928 S1 When two IOs attempt to access the same address, the state of the data may be incorrectly set to invalid causing offline volumes and, possibly, offline pools 8.2.1.4 Data Reduction Pools
HU01987 S1 During SAN fabric power maintenance an AC3 cluster may breech resource limits, on the remaining node to node links, resulting in system-wide lease expiry 8.2.1.4 Reliability Availability Serviceability
HU02000 S1 Data Reduction Pools may go offline due to a timing issue in metadata handling 8.2.1.4 Data Reduction Pools
HU02013 S1 A race condition, between the extent invalidation and destruction, in the garbage collection process, may cause a node warmstart with the possibility of offline volumes 8.2.1.4 Data Reduction Pools
HU02025 S1 An issue with metadata handling, where a pool has been taken offline, may lead to an out of space condition, in that pool, preventing its return to operation 8.2.1.4 Data Reduction Pools
HU01886 S2 The Unmap function can leave volume extents, that have not been freed, preventing managed disk and pool removal 8.2.1.4 SCSI Unmap
HU01941 S2 After upgrading the system to v8.2, or later, when expanding a mirrored volume, the format of additional space may become stalled 8.2.1.4 Volume Mirroring
HU01944 S2 Proactive host failover not waiting for 25 seconds before allowing nodes to go offline during upgrades or maintenance 8.2.1.4 Reliability Availability Serviceability
HU01972 S2 When an array is in a quiescing state, for example where a member has been deleted, I/O may become pended leading to multiple warmstarts 8.2.1.4 RAID, Distributed RAID
HU00744 S3 Single node warmstart due to an accounting issue within the cache component 8.2.1.4 Cache
HU01485 S3 When a AC3 node is started, with only one PSU powered, powering up the other PSU will not extinguish the Power Fault LED.
Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed
8.2.1.4 System Monitoring
HU01659 S3 On AC3 systems the Node Fault LED may be seen to flash in the absence of an error condition.
Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed
8.2.1.4 System Monitoring
HU01857 S3 Improved validation of user input in GUI 8.2.1.4 Graphical User Interface
HU01860 S3 During garbage collection the flushing of extents may become stuck leading to a timeout and a single node warmstart 8.2.1.4 Data Reduction Pools
HU01869 S3 Volume copy deletion in a Data Reduction Pool triggered by rmvdiskcopy rmvolumecopy or addvdiskcopy -autodelete (or similar) may become stalled with the copy being left in "deleting" status 8.2.1.4 Data Reduction Pools
HU01911 S3 The "System Overview" screen, in the GUI, may display nodes in the wrong site 8.2.1.4 Graphical User Interface
HU01912 S3 Systems with iSCSI-attached controllers may see node warmstarts due to I/O request timeouts 8.2.1.4 Backend Storage
HU01915 & IT28654 S3 For systems, with encryption enabled, that are using key servers to manage encryption keys, the system may fail to connect to the key servers if the server's SSL certificates is part of a chain of trust 8.2.1.4 Encryption
HU01916 S3 The GUI Dashboard and the CLI lssystem command report physical capacity incorrectly 8.2.1.4 Graphical User Interface, Command Line Interface
HU01926 S3 When a node with 32GB of RAM is upgraded to v8.2.1 it may experience a warmstart resulting in a failed upgrade 8.2.1.4 System Update
HU01959 S3 An timing window issue in the Thin Provisioning component can cause a node warmstart 8.2.1.4 FlashCopy, Thin Provisioning
HU01962 S3 When Call Home servers return an invalid message it can be incorrectly reported as an error 3201 in the Event Log 8.2.1.4 System Monitoring
HU01976 S3 A new MDisk array may not be encrypted even though encryption is enabled on the system 8.2.1.4 Encryption
HU02001 S3 During a system upgrade an issue in callhome may cause a node warmstart, stalling the upgrade 8.2.1.4 System Monitoring
HU02002 S3 On busy systems, diagnostic data collection may not complete correctly producing livedumps with missing pages 8.2.1.4 Support Data Collection
HU02019 S3 When the master and auxiliary volumes, in a relationship, have the same name it is not possible, in the GUI, to determine which is master or auxiliary 8.2.1.4 Graphical User Interface
HU02166 S3 A timing window issue, in RAID code that handles recovery after a drive has been taken out of sync, due to a slow I/O, can cause a single node warmstart 8.2.1.4 RAID
IT28433 S3 Timing window issue in the Data Reduction Pool rehoming component can cause a single node warmstart 8.2.1.4 Data Reduction Pools
IT28728 S3 Email alerts will not work where the mail server does not allow unqualified client host names 8.2.1.4 System Monitoring
HU01932 S1 When a rmvdisk command initiates a Data Reduction Pool rehoming process, any I/O to deduplicated data, referenced to data on the removed volume, may cause multiple warmstarts leading to a loss of access. 8.2.1.2 Deduplication
FLASH-28023 S2 V9000 no longer sending heartbeats. 8.2.1.2 Reliability Availability Serviceability
HU01920 S1 An issue in the garbage collection process can cause node warmstarts and offline pools 8.2.1.1 Data Reduction Pools
HU01492 & HU02024 S1 HIPER (Highly Pervasive): All ports of a 16Gb HBA can be affected when a single port is congested. This can lead to lease expiries if all ports used for inter-node communication are on the same FC adapter 8.2.1.0 Reliability Availability Serviceability
HU01617 S1 HIPER (Highly Pervasive): Due to a timing window issue, stopping a FlashCopy mapping, with the -autodelete option, may result in a Tier 2 recovery 8.2.1.0 FlashCopy
HU01873 S1 HIPER (Highly Pervasive): Deleting a volume, in a Data Reduction Pool, while volume protection is enabled and when the volume was not explicitly unmapped, before deletion, may result in simultaneous node warmstarts. For more details refer to the following Flash 8.2.1.0 Data Reduction Pools
HU01743 S1 Where hosts are directly attached, a mishandling of the login process, by the fabric controller, may result in dual node warmstarts 8.2.1.0 Hosts
HU01758 S1 After an unexpected power loss, all nodes, in a cluster, may warmstart repeatedly, necessitating a Tier 3 recovery 8.2.1.0 RAID
HU01799 S1 Timing window issue can affect operation of the HyperSwap addvolumecopy command causing all nodes to warmstart 8.2.1.0 HyperSwap
HU01825 S1 Invoking a chrcrelationship command when one of the relationships, in a consistency group, is running in the opposite direction to the others may cause a node warmstart followed by a Tier 2 recovery 8.2.1.0 FlashCopy
HU01828 S1 Node warmstarts may occur during deletion of deduplicated volumes due to a timing-related issue 8.2.1.0 Deduplication
HU01833 S1 If both nodes, in an I/O group, start up together a timing window issue may occur, that would prevent them running garbage collection, leading to a related Data Reduction Pool running out of space 8.2.1.0 Data Reduction Pools
HU01845 S1 If the execution of a rmvdisk -force command, for the FlashCopy target volume in a GMCV relationship, coincides with the start of a GMCV cycle all nodes may warmstart 8.2.1.0 Global Mirror with Change Volumes
HU01847 S1 FlashCopy handling of medium errors, across a number of drives on backend controllers, may lead to multiple node warmstarts 8.2.1.0 FlashCopy
HU01848 S1 During an upgrade, systems with a large AIX VIOS setup may have multiple node warmstarts with the possibility of a loss of access to data 8.2.1.0 System Update
HU01850 S1 When the last deduplication-enabled volume copy, in a Data Reduction Pool, is deleted, the pool may go offline temporarily 8.2.1.0 Data Reduction Pools, Deduplication
HU01851 S1 When a deduplicated volume is deleted there may be multiple node warmstarts and offline pools 8.2.1.0 Deduplication
HU01855 S1 Clusters using Data Reduction Pools can experience multiple warmstarts on all nodes putting them in a service state 8.2.1.0 Data Reduction Pools
HU01862 S1 When a Data Reduction Pool is removed, and the -force option is specified, there may be a temporary loss of access 8.2.1.0 Data Reduction Pools
HU01876 S1 Where systems are connected to controllers, that have FC ports that are capable of acting as initiators and targets, when NPIV is enabled then node warmstarts can occur 8.2.1.0 Backend Storage
HU01878 S1 During an upgrade, from v7.8.1 or earlier to v8.1.3 or later, if an MDisk goes offline then, at completion, all volumes may go offline 8.2.1.0 System Update
HU01885 S1 As writes are made to a Data Reduction Pool it is necessary to allocate new physical capacity. Under unusual circumstances it is possible for the handling of an expansion request to stall further I/O leading to node warmstarts 8.2.1.0 Data Reduction Pools
HU01913 S1 A timing window issue in the DRAID6 rebuild process can cause node warmstarts with the possibility of a loss of access 8.2.1.0 Distributed RAID
HU01957 S1 Due to an issue in Data Reduction Pools, when the system attempts an upgrade, there may be node warmstarts 8.2.1.0 Data Reduction Pools, System Update
HU01965 S1 A timing window issue in the deduplication component can lead to I/O timeouts, and a node warmstart, with the possibility of an offline mdisk group 8.2.1.0 Deduplication
HU02042 S1 An issue in the handling of metadata, after a Data Reduction Pool recovery operation, can lead to repeated node warmstarts, putting an I/O group into a service state 8.2.1.0 Data Reduction Pools
IT25850 S1 I/O performance may be adversely affected towards the end of DRAID rebuilds. For some systems there may be multiple warmstarts leading to a loss of access 8.2.1.0 Distributed RAID
IT27460 S1 Lease expiry can occur between local nodes when remote connection is lost, due to the mishandling of messaging credits 8.2.1.0 Reliability Availability Serviceability
IT29040 S1 Occasionally a DRAID rebuild, with drives of 8TB or more, can encounter an issue which causes node warmstarts and potential loss of access 8.2.1.0 RAID, Distributed RAID
FLASH-27506 S2 Improved RAID error handling for unresponsive flash modules to prevent rare data error 8.2.1.0 RAID
HU01507 S2 Until the initial synchronisation process completes, high system latency may be experienced when a volume is created with two compressed copies or when space-efficient copy is added to a volume with an existing compressed copy 8.2.1.0 Volume Mirroring
HU01661 S2 A cache-protection mechanism flag setting can become stuck, leading to repeated stops of consistency group synching 8.2.1.0 HyperSwap
HU01733 S2 Canister information, for the High Density Expansion Enclosure, may be incorrectly reported 8.2.1.0 Reliability Availability Serviceability
HU01761 S2 Entering multiple addmdisk commands, in rapid succession, to more than one storage pool, may cause node warmstarts 8.2.1.0 Backend Storage
HU01797 S2 Hitachi G1500 backend controllers may exhibit higher than expected latency 8.2.1.0 Backend Storage
HU01810 S2 Deleting volumes, or using FlashCopy/Global Mirror with Change Volumes, in a Data Reduction Pool, may impact the performance of other volumes in the pool. 8.2.1.0 Data Reduction Pools
HU01837 S2 In systems, where a VVols metadata volume has been created, an upgrade to v8.1.3 or later will cause a node warmstart, stalling the upgrade 8.2.1.0 VVols
HU01839 S2 Where a VMware host is being served volumes, from two different controllers, and an issue, on one controller, causes the related volumes to be taken offline then I/O performance, for the volumes from the other controller, will be adversely affected 8.2.1.0 Hosts
HU01842 S2 Bursts of I/O to Samsung Read-Intensive Drives can be interpreted as dropped frames, against the resident slots, leading to redundant drives being incorrectly failed 8.2.1.0 Drives
HU01846 S2 Silent battery discharge condition will, unexpectedly, take a SVC node offline, putting it into a 572 service state 8.2.1.0 Reliability Availability Serviceability
HU01852 S2 The garbage collection rate can lead to Data Reduction Pools running out of space even though reclaimable capacity is available 8.2.1.0 Data Reduction Pools
HU01858 S2 Total used capacity of a Data Reduction Pool within a single I/O group is limited to 256TB. Garbage collection does not correctly recognise this limit. This may lead to a pool running out of free capacity and going offline 8.2.1.0 Data Reduction Pools
HU01881 S2 An issue within the compression card in FS9100 systems can result in the card being incorrectly flagged as failed leading to warmstarts 8.2.1.0 Compression
HU01883 S2 Config node processes may consume all available memory, leading to node warmstarts. This can be caused, for example, by large numbers of concurrent SSH connections being opened 8.2.1.0 Reliability Availability Serviceability
HU01907 S2 An issue in the handling of the power cable sense registers can cause a node to be put into service state with a 560 error 8.2.1.0 Reliability Availability Serviceability
HU02076 S2 For systems with AE2 enclosures, error 1007 (Fewer bus ports operational) should generate a callhome 8.2.1.0 System Monitoring
HU00921 S3 A node warmstart may occur when an MDisk state change gives rise to duplicate discovery processes 8.2.1.0
HU01276 S3 An issue in the handling of debug data from the FC adapter can cause a node warmstart 8.2.1.0 Reliability Availability Serviceability
HU01523 S3 An issue with FC adapter initialisation can lead to a node warmstart 8.2.1.0 Reliability Availability Serviceability
HU01564 S3 FlashCopy maps cleaning process is not monitoring the grains correctly which may cause FlashCopy maps to not stop 8.2.1.0 FlashCopy
HU01571 S3 An upgrade can become stalled due to a node warmstart 8.2.1.0 System Update
HU01657 S3 The 16Gb FC HBA firmware may experience an issue, with the detection of unresponsive links, leading to a single node warmstart 8.2.1.0 Reliability Availability Serviceability
HU01667 S3 A timing-window issue, in the remote copy component, may cause a node warmstart 8.2.1.0 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01719 S3 Node warmstart due to a parity error in the HBA driver firmware 8.2.1.0 Reliability Availability Serviceability
HU01737 S3 On the "Update System" screen, for "Test Only", if a valid code image is selected, in the "Run Update Test Utility" dialog, then clicking the "Test" button will initiate a system update 8.2.1.0 System Update
HU01751 S3 When RAID attempts to flag a strip as bad, and that strip has already been flagged, a node may warmstart 8.2.1.0 RAID
HU01760 S3 FlashCopy map progress appears to be stuck at zero percent 8.2.1.0 FlashCopy
HU01765 S3 Node warmstart may occur when there is a delay to I/O at the secondary site 8.2.1.0 Global Mirror, Global Mirror with Change Volumes, Metro Mirror
HU01772 S3 The mail queue may become blocked preventing the transmission of event log messages 8.2.1.0 System Monitoring
HU01784 S3 If a cluster using IP quorum experiences a site outage, the IP quorum device may become invalid. Restarting the quorum application will resolve the issue 8.2.1.0 HyperSwap, Quorum
HU01786 S3 An issue in the monitoring of SSD write endurance can result in false 1215/2560 errors in the Event Log 8.2.1.0 Drives
HU01791 S3 Using the chhost command will remove stored CHAP secrets 8.2.1.0 iSCSI
HU01807 S3 The lsfabric command may show incorrect local node id and local node name for some Fibre Channel logins 8.2.1.0 Command Line Interface
HU01811 S3 DRAID rebuilds, for large (>10TB) drives, may require lengthy metadata processing leading to a node warmstart 8.2.1.0 Distributed RAID
HU01815 S3 In Data Reduction Pools, volume size is limited to 96TB 8.2.1.0 Data Reduction Pools
HU01817 S3 Volumes used for VVols metadata or cloud backup, that are associated with a FlashCopy mapping, cannot be included in any further FlashCopy mappings 8.2.1.0 FlashCopy
HU01821 S3 An attempt to upgrade a two-node enhanced stretched cluster fails due to incorrect volume dependencies 8.2.1.0 System Update, Data Reduction Pools
HU01832 S3 Creation and distribution of the config file may cause an out-of-memory condition, leading to a node warmstart 8.2.1.0 Reliability Availability Serviceability
HU01849 S3 An excessive number of SSH sessions may lead to a node warmstart 8.2.1.0 System Monitoring
HU01856 S3 A garbage collection process can time out waiting for an event in the partner node resulting in a node warmstart 8.2.1.0 Data Reduction Pools
HU01863 S3 In rare circumstances, a drive replacement may result in a "ghost drive" (i.e. a drive with the same ID as the replaced drive stuck in a permanently offline state) 8.2.1.0 Drives
HU01871 S3 An issue with bitmap synchronisation can lead to a node warmstart 8.2.1.0 Data Reduction Pools
HU01879 S3 Latency induced by DWDM inter-site links may result in a node warmstart 8.2.1.0
HU01893 S3 Excessive reporting frequency of NVMe drive diagnostics generates large numbers of callhome events 8.2.1.0 Drives
HU01895 S3 Where a banner has been created, without a new line at the end, any subsequent T4 recovery will fail 8.2.1.0 Distributed RAID
HU01981 S3 Although an issue in the 16Gb HBA firmware is handled correctly it can still cause a node warmstart 8.2.1.0 Reliability Availability Serviceability
HU02028 S3 An issue, with timer cancellation, in the Remote Copy component may cause a node warmstart 8.2.1.0 Metro Mirror, Global Mirror, Global Mirror with Change Volumes
IT19561 S3 An issue with register clearance in the FC driver code may cause a node warmstart 8.2.1.0 Reliability Availability Serviceability
IT25457 S3 Attempting to remove a copy of a volume, which has at least one image mode copy and at least one thin/compressed copy, in a Data Reduction Pool will always fail with a CMMVC8971E error 8.2.1.0 Data Reduction Pools
IT25970 S3 After a FlashCopy consistency group is started a node may warmstart 8.2.1.0 FlashCopy
IT26049 S3 An issue with CPU scheduling may cause the GUI to respond slowly 8.2.1.0 Graphical User Interface

4. Supported upgrade paths

Please refer to the Concurrent Compatibility and Code Cross Reference for Spectrum Virtualize page for guidance when planning a system upgrade.

5. Useful Links

Description Link
Support Website IBM Knowledge Center
IBM FlashSystem Fix Central V9000
Updating the system IBM Knowledge Center
IBM Redbooks Redbooks
Contacts IBM Planetwide