This document was last updated on 7th February 2023.
Details | Introduced |
---|---|
Customers with more than 5 x non-NVMe over FC hosts (i.e FC SCSI or iSCSI) in an I/O group must not attach any NVMe over FC hosts to that I/O group. These limits will not be policed by the Spectrum Virtualize software. Any configurations that exceed these limits will experience significant adverse performance impact. These limits will be lifted in a future major release. |
8.2.1.0 |
Customers using Transparent Cloud Tiering should not upgrade to v8.2.1.0. This is a restriction that may be lifted in a future PTF. |
8.2.1.0 |
Customers using iSCSI to virtualize backend controllers should not upgrade to v8.2.0 or later This is a restriction that may be lifted in a future PTF. |
8.2.0.0 |
Customers using the REST API to list more than 2000 objects may experience a loss of service from the API as it restarts due to memory constraints. This is a restriction that may be lifted in a future PTF. |
8.1.3.0 |
It is not possible to access the REST API using a cluster's IPv6 address. This is a restriction that may be lifted in a future PTF. |
8.1.3.0 |
Customers upgrading systems with more than 64GB of RAM to v8.1 or later will need to run chnodehw to enable access to the extra memory above 64GB. Under some circumstances it may also be necessary to remove and re-add each node in turn. |
8.1.0.0 |
Validation in the Upload Support Package feature will reject the new case number format in the PMR field. This is a known issue that may be lifted in a future PTF. The fix can be tracked using APAR HU02392. |
7.8.1.0 |
Systems, with NPIV enabled, presenting storage to SUSE Linux Enterprise Server (SLES) or Red Hat Enterprise Linux (RHEL) hosts running the ibmvfc driver on IBM Power can experience path loss or read-only file system events. This is cause by issues within the ibmvfc driver and VIOS code. Refer to this troubleshooting page for more information. |
n/a |
If an update stalls or fails then contact IBM Support for further assistance | n/a |
The following restrictions were valid but have now been lifted | |
Customers with direct attached external storage controllers cannot upgrade to v8.2.1.6. This is most likely to impact customers with 2 x1 Direct Attach V9000 systems. This has been resolved, under APAR HU02077, in v8.2.1.8. Please note that v8.2.1.5, or earlier, is not exposed to this restriction. |
8.2.1.6 |
Customers using systems with attached AE2 enclosures cannot upgrade to versions 8.2.1.0 - 8.2.1.4, but rather, must upgrade directly to v8.2.1.5 or later. |
8.2.1.0 |
With Gemalto SafeNet KeySecure, the chkeyserverkeysecure -username <username> command is used to set the KeySecure username credential. If this is changed to a username that is not recognised by the key server to be the valid username, associated with the Spectrum Virtualize encryption key, then a subsequent re-key operation can cause key servers to appear offline. This is a an issue that will be resolved in a future PTF. |
8.2.1.0 |
A release may contain fixes for security issues, fixes for APARs or both. Consult both tables below to understand the complete set of fixes included in the release.
CVE Identifier | Link for additional Information | Resolved in |
---|---|---|
CVE-2023-2597 | 7065011 | 8.2.1.18 |
CVE-2022-21626 | 6858041 | 8.2.1.17 |
CVE-2022-43873 | 6858047 | 8.2.1.17 |
CVE-2022-0778 | 6622017 | 8.2.1.16 |
CVE-2021-35603 | 6622019 | 8.2.1.16 |
CVE-2021-35550 | 6622019 | 8.2.1.16 |
CVE-2018-25032 | 6622021 | 8.2.1.16 |
CVE-2021-38969 | 6584337 | 8.2.1.15 |
CVE-2021-29873 | 6497111 | 8.2.1.14 |
CVE-2020-2781 | 6445063 | 8.2.1.12 |
CVE-2020-13935 | 6445063 | 8.2.1.12 |
CVE-2020-14577 | 6445063 | 8.2.1.12 |
CVE-2020-14578 | 6445063 | 8.2.1.12 |
CVE-2020-14579 | 6445063 | 8.2.1.12 |
CVE-2019-5544 | 6250889 | 8.2.1.11 |
CVE-2019-2964 | 6250887 | 8.2.1.11 |
CVE-2019-2989 | 6250887 | 8.2.1.11 |
CVE-2018-12404 | 6250885 | 8.2.1.11 |
CVE-2019-11477 | 1164286 | 8.2.1.6 |
CVE-2019-11478 | 1164286 | 8.2.1.6 |
CVE-2019-11479 | 1164286 | 8.2.1.6 |
CVE-2019-2602 | 1073958 | 8.2.1.6 |
CVE-2018-3180 | ibm10884526 | 8.2.1.4 |
CVE-2018-12547 | ibm10884526 | 8.2.1.4 |
CVE-2008-5161 | ibm10874368 | 8.2.1.2 |
CVE-2018-5391 | ibm10872368 | 8.2.1.2 |
CVE-2018-11776 | ibm10741137 | 8.2.1.0 |
CVE-2017-17449 | ibm10872364 | 8.2.1.0 |
CVE-2017-18017 | ibm10872364 | 8.2.1.0 |
CVE-2017-17833 | ibm10872546 | 8.2.1.0 |
CVE-2018-11784 | ibm10872550 | 8.2.1.0 |
CVE-2018-5732 | ibm10741135 | 8.2.1.0 |
CVE-2018-1517 | ibm10872456 | 8.2.1.0 |
CVE-2018-2783 | ibm10872456 | 8.2.1.0 |
CVE-2018-12539 | ibm10872456 | 8.2.1.0 |
CVE-2018-1775 | ibm10872486 | 8.2.1.0 |
Reference | Severity | Description | Resolved in | Feature Tags |
---|---|---|---|---|
HU02327 | S1 | Using addvdiskcopy in conjunction with expandvdisk with format may result in the original being overwritten, by the new copy, producing blank copies. For more details refer to the following Flash | 8.2.1.15 | Volume Mirroring |
HU02400 | S1 | A problem in the virtualization component of the system can cause a migration IO to be submitted in an incorrect context resulting in a node warmstart. In some cases it is possible that this IO has been submitted to an incorrect location on the backend, which can cause data corruption of an isolated small area | 8.2.1.15 | Storage Virtualisation |
HU02342 | S1 | Occasionally when an offline drive returns to online state later than its peers in the same RAID array there can be multiple node warmstarts that send nodes into a service state | 8.2.1.15 | RAID |
HU02374 | S1 | Hosts with Emulex 16Gbps HBAs may become unable to communicate with a system with 8Gbps Fibre Channel ports, after the host HBA is upgraded to firmware version 12.8.364.11. This does not apply to systems with 16Gb or 32Gb Fibre Channel ports | 8.2.1.15 | Hosts |
HU02393 | S1 | Automatic resize of compressed/thin volumes may fail causing warmstarts on both nodes in an I/O group | 8.2.1.15 | Storage Virtualisation |
HU02401 | S1 | EasyTier can move extents between identical mdisks until one runs out of space | 8.2.1.15 | EasyTier |
HU02406 | S1 | An interoperability issue between Cisco NX-OS firmware and the Spectrum Virtualize Fibre Channel driver can cause a node warmstart on NPIV failback (for example during an upgrade) with the potential for a loss of access. For more details refer to the following Flash | 8.2.1.15 | Interoperability |
IT38015 | S2 | During RAID rebuild or copyback on systems with 16gb or less of memory, cache handling can lead to a deadlock which results in timeouts | 8.2.1.15 | RAID |
HU02366 | S3 | Slow internal resource reclamation by the RAID component can cause a node warmstart | 8.2.1.15 | RAID |
HU02433 | S3 | When a BIOS upgrade occurs excessive tracefile entries can be generated | 8.2.1.15 | System Update |
HU02527 | S3 | If more than 15 SAS enclosures are attached, callhome may fail due to the email being too large for the server to process. | 8.2.1.15 | System Monitoring |
HU02277 | S1 | HIPER (Highly Pervasive): RAID parity scrubbing can become stalled causing an accumulation of media errors leading to multiple drive failures with the possibility of data integrity loss. For more details refer to the following Flash | 8.2.1.12 | RAID |
HU01968 & HU02215 | S1 | An upgrade may fail due to corrupt hardened data in a node. This can affect an I/O group | 8.2.1.12 | System Update |
HU02058 | S1 | Changing a remote copy relationship from GMCV to MM or GM can result in a Tier 2 recovery | 8.2.1.12 | Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU02213 | S1 | A Hot Spare Node (HSN) timing window issue can, during an HSN activation or deactivation, cause the cluster to broadcast an invalid VPD update to other clusters on the SAN. This may trigger a Tier 2 recovery on the other cluster. For more details refer to the following Flash | 8.2.1.12 | Hot Spare Node |
HU02266 | S1 | An issue in auto-expand can cause expansion to fail and the volume to be taken offline | 8.2.1.12 | Thin Provisioning |
HU02295 | S1 | When upgrading from v8.2.1 or v8.3, in the presence of hot spare nodes, an issue with the handling of node metadata may cause a Tier 2 recovery | 8.2.1.12 | System Update |
HU02429 | S1 | System can go offline shortly after changing the SMTP settings using the chemailserver command via the GUI | 8.2.1.12 | System Monitoring |
HU02132 | S2 | Removing a thin-provisioned volume and then immediately creating one of the same size may cause node warmstarts | 8.2.1.12 | Thin Provisioning |
HU02156 | S2 | Global Mirror environments may experience more frequent 1920 events due to writedone message queuing | 8.2.1.12 | Global Mirror |
HU02164 | S2 | An issue in Remote Copy may cause a loss of hardened data when a node is warmstarted | 8.2.1.12 | Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU02176 | S2 | During upgrade a node may limit the number of target ports it reports causing a failover contradiction on hosts | 8.2.1.12 | Hosts |
HU02184 | S1 | When a 3PAR controller experiences a fault that prevents normal I/O processing it may issue a SCSI TARGET RESET command. This command is not supported and may cause multiple node asserts, possibly cluster-wide | 8.2.1.12 | Backend Storage |
HU02200 | S2 | When upgrading from v8.1 or earlier to v8.2.1 or later a remote copy issue may cause a node warmstart, stalling the upgrade | 8.2.1.12 | System Update |
HU02201 & HU02221 | S2 | Shortly after upgrading drive firmware, specific drive models can fail due to Too many long IOs to drive for too longerrors |
8.2.1.12 | Drives |
HU02227 | S2 | Certain I/O patterns can cause compression hardware to post errors. When those errors exceed a threshold the node can be taken offline | 8.2.1.12 | Compression |
HU02288 | S2 | A node might fail to come online after a reboot or warmstart such as during an upgrade | 8.2.1.12 | Reliability Availability Serviceability |
HU02048 | S3 | An issue in the handling of ATS commands from VMware hosts can cause a single node warmstart | 8.2.1.12 | Hosts |
HU02142 | S3 | It is possible for a backend unmap process to become stalled, preventing system configuration changes from completing | 8.2.1.12 | Distributed RAID |
HU02157 | S3 | Issuing a mkdistributedarray command may result in a node warmstart | 8.2.1.12 | Distributed RAID |
HU02241 | S3 | IP Replication can fail to create IP partnerships via the secondary cluster management IP | 8.2.1.12 | IP Replication |
HU02244 | S3 | False positive node error 766 (depleted CMOS battery) messages may appear in the Event Log | 8.2.1.12 | System Monitoring |
HU02292 & HU02308 | S3 | The use of maximum replication delay within Global Mirror may occasionally cause a node warmstart | 8.2.1.12 | Global Mirror |
HU02332 & HU02336 | S3 | When an I/O is received, from a host, with invalid or inconsistent SCSI data but a good checksum it may cause a node warmstart | 8.2.1.12 | Hosts |
HU02354 | S3 | An issue in the handling of read transfers may cause hung host IOs leading to a node warmstart | 8.2.1.12 | Hosts |
HU02358 | S3 | An issue in Remote Copy, that stalls a switch of direction, can cause I/O timeouts leading to a node warmstart | 8.2.1.12 | Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01894 | S1 | HIPER (Highly Pervasive): After node reboot, or warmstart, some volumes accessed by AIX, VIO or VMware hosts, may experience stuck SCSI2 reservations on the NPIV failover ports of the partner node. This can cause a loss of access to data | 8.2.1.11 | Hosts |
HU02141 | S1 | HIPER (Highly Pervasive): An issue in the max replication delay function may trigger a Tier 2 recovery, after posting multiple 1920 errors in the Event Log. For more details refer to the following Flash | 8.2.1.11 | Global Mirror |
HU02205 | S1 | HIPER (Highly Pervasive): Incremental FlashCopy targets can be corrupted when the FlashCopy source is a target of a remote copy relationship | 8.2.1.11 | FlashCopy, Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU02212 | S1 | HIPER (Highly Pervasive): Remote Copy secondary may have inconsistent data following a stop with -access due to a missing bitmap merge from FlashCopy to Remote Copy For more details refer to the following Flash | 8.2.1.11 | Global Mirror with Change Volumes, HyperSwap |
HU02237 | S1 | HIPER (Highly Pervasive): Under a rare and complicated set of conditions, a RAID 1 or RAID 10 array may drop a write, causing undetected data corruption. For more details refer to the following Flash | 8.2.1.11 | RAID |
HU02238 | S1 | HIPER (Highly Pervasive): Force-stopping a FlashCopy map, where the source volume is a Metro or Global Mirror target volume, may cause other FlashCopy maps to return invalid data if they are not 100% copied, in specific configurations. For more details refer to the following Flash | 8.2.1.11 | RAID |
HU01921 | S1 | Where FlashCopy mapping targets are also in remote copy relationships there may be node warmstarts with a temporary loss of access to data | 8.2.1.11 | FlashCopy, Global Mirror, Metro Mirror |
HU01924 | S1 | Migrating extents to an MDisk, that is not a member of an MDisk group, may result in a Tier 2 recovery | 8.2.1.11 | Thin Provisioning |
HU01970 | S1 | When a GMCV relationship is stopped with the -access option, and the secondary volume is immediately deleted with -force, then all nodes may repeatedly warmstart | 8.2.1.11 | Global Mirror with Change Volumes |
HU02005 | S1 | An issue in the background copy process prevents grains, above a 128TB limit, from being cleaned properly. As a consequence there may be multiple node warmstarts with the potential for a loss of access to data | 8.2.1.11 | Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU02054 | S1 | The event log handler maintains a second list of events. On rare occasions, for log full events, these lists can get out of step, resulting in a Tier 2 recovery | 8.2.1.11 | System Monitoring |
HU02065 | S1 | Mishandling of Data Reduction Pool allocation request rejections can lead to node warmstarts that can take an MDisk group offline | 8.2.1.11 | Data Reduction Pools |
HU02069 | S1 | When a SCSI command, containing an invalid byte, is received there may be a node warmstart. This can affect both nodes, in an I/O group, at the same time | 8.2.1.11 | Hosts |
HU02097 | S1 | Workloads, with data that is highly suited to deduplication, can provoke high CPU utilisation, as multiple destinations try to dedupe to one source. This adversely impacts performance with the possibility of offline MDisk groups | 8.2.1.11 | Data Reduction Pools |
HU02106 | S1 | Multiple node warmstarts, in quick succession, can cause the partner node to lease expire | 8.2.1.11 | IP Quorum |
HU02108 | S1 | Deleting a managed disk group, with -force, may cause multiple warmstarts with the possibility of a loss of access to data | 8.2.1.11 | Data Reduction Pools |
HU02109 | S1 | Free extents may not be unmapped after volume deletion, or migration, resulting in out-of-space conditions on backend controllers | 8.2.1.11 | Backend Storage, SCSI Unmap |
HU02135 | S1 | Removing multiple IQNs for an iSCSI host can result in a Tier 2 recovery | 8.2.1.11 | iSCSI |
HU02138 | S1 | An issue in Data Reduction Pool garbage collection can cause I/O timeouts leading to an offline pool | 8.2.1.11 | Data Reduction Pools |
HU02154 | S1 | If a node is rebooted, when remote support is enabled, then all other nodes will warmstart | 8.2.1.11 | Support Remote Assist |
HU02155 | S1 | Upgrading to v8.2.1 may result in offline managed disk groups and OOS events (1685/1687) appearing in the Event Log | 8.2.1.11 | Data Reduction Pools |
HU02197 | S1 | Bulk volume removals can adversely impact related FlashCopy mappings leading to a Tier 2 recovery | 8.2.1.11 | FlashCopy |
HU02222 | S1 | Where the source volume of an incremental FlashCopy map is also a Metro or Global Mirror target volume that is using a change volume or is a Hyperswap volume, then there is a possibility that not all data will be copied to the FlashCopy target. For more details refer to the following Flash | 8.2.1.11 | Global Mirror with Change Volumes |
IT25367 | S1 | A T2 recovery may occur when an attempt is made to upgrade, or downgrade, the firmware for an unsupported drive type | 8.2.1.11 | Drives |
IT31113 | S1 | After a manual power off and on, of a system, both nodes, in an I/O group, may repeatedly assert into a service state | 8.2.1.11 | RAID |
HU01923 | S2 | An issue in the way Global Mirror handles write sequence numbers >512 may cause multiple node warmstarts | 8.2.1.11 | Global Mirror |
HU02080 | S2 | When a Data Reduction Pool is running low on free space, the credit allocation algorithm, for garbage collection, can be exposed to a race condition, adversely affecting performance | 8.2.1.11 | Data Reduction Pools |
HU02123 | S2 | For direct-attached hosts, a race condition between the FLOGI and Link UP processes can result in FC ports not coming online | 8.2.1.11 | Hosts |
HU02149 | S2 | When an Enhanced Stretch Cluster is using NPIV, in transitional mode, the path priority is not being reported correctly to some hosts | 8.2.1.11 | Hosts |
HU01868 | S3 | After deleting an encrypted external MDisk, it is possible for the 'encrypted' status of volumes to change to 'no', even though all remaining MDisks are encrypted | 8.2.1.11 | Encryption |
HU01917 | S3 | Chrome browser support requires a self-signed certificate to include subject alternate name | 8.2.1.11 | Graphical User Interface |
HU01931 | S3 | Where a high rate of CLI commands are received, it is possible for inter-node processing code to be delayed which results in a small increase in receive queue time on the config node | 8.2.1.11 | Performance |
HU02111 | S3 | An issue with how Data Reduction Pools handle data, at the sub-extent level, may result in a node warmstart | 8.2.1.11 | Data Reduction Pools |
HU02124 | S3 | Due to an issue with FCM thin provisioning calculations the GUI may incorrectly display volume capacity and capacity savings as zero | 8.2.1.11 | System Monitoring |
HU02137 | S3 | An issue with support for target resets in Nimble Storage controllers may cause a node warmstart | 8.2.1.11 | Backend Storage |
HU02173 | S3 | During a pending fabric login, when an abort is received, it is possible for a related entry in the WWPN table to not be removed. The node will warmstart to clear this condition. | 8.2.1.11 | Reliability Availability Serviceability |
HU02183 | S3 | An issue in the way inter-node communication is handled can lead to a node warmstart | 8.2.1.11 | Reliability Availability Serviceability |
HU02190 | S3 | Error 1046 not triggering a Call Home even though it is a hardware fault | 8.2.1.11 | System Monitoring |
HU02214 | S3 | Under a certain I/O pattern it is possible for metadata management in Data Reduction Pools to become inconsistent leading to a node warmstart | 8.2.1.11 | Data Reduction Pools |
HU02143 | S2 | The performance profile, for some enterprise tier drives, may not correctly match the drives capabilities leading to that tier being overdriven | 8.2.1.10 | EasyTier |
HU02104 | S1 | HIPER (Highly Pervasive): An issue in the RAID component, in the presence of very high I/O workload and the exhaustion of cache resources, can see a deadlock condition occurring which prevents further I/O processing. The system detects this issue and takes the storage pool offline for a six minute period, to clear the problem. The pool is then brought online automatically, and normal operation resumes. For more details refer to the following Flash | 8.2.1.9 | RAID, Distributed RAID |
HU02102 | S3 | Excessive processing time required for FlashCopy bitmap operations, associated with large (>20TB) Global Mirror change volumes, may lead to a node warmstart | 8.2.1.9 | Global Mirror with Change Volumes |
HU02131 | S3 | When changing DRAID configuration, for an array with an active workload, a deadlock condition can occur resulting in a single node warmstart | 8.2.1.9 | Distributed RAID |
HU02126 | S3 | There is a low probability that excessive SSH connections may trigger a single node warmstart on the configuration node | 8.2.1.9 | Command Line Interface |
HU02064 | S1 | HIPER (Highly Pervasive): An issue in the firmware for compression accelerator cards can cause offline compressed volumes. For more details refer to the following Flash | 8.2.1.8 | Compression |
HU02083 | S1 | HIPER (Highly Pervasive): During DRAID rebuilds, an issue in the handling of memory buffers can lead to multiple node warmstarts and a loss of access to data For more details refer to the following Flash | 8.2.1.8 | Distributed RAID |
HU01967 | S1 | When I/O, in remote copy relationships, experiences delays (1720 and/or 1920 errors are logged) an I/O group may warmstart | 8.2.1.8 | Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU02036 | S1 | It is possible for commands, that alter pool-level extent reservations (i.e. migratevdisk or rmmdisk), to conflict with an ongoing EasyTier migration, resulting in a Tier 2 recovery | 8.2.1.8 | EasyTier |
HU02044 | S1 | Multiple DRAID arrays can, where one is performing a rebuild, be exposed to a RAID deadlock condition resulting in multiple node warmstarts and a loss of access to data | 8.2.1.8 | Distributed RAID |
HU02063 | S1 | HyperSwap clusters with only two surviving nodes may experience warmstarts on both of those nodes where rcbuffersize is set to 512MB | 8.2.1.8 | HyperSwap |
HU02077 | S1 | Upgrading to 8.2.1.6 or 8.3.0.0 can cause a loss of access to direct-attached Fibre Channel controllers | 8.2.1.8 | Backend Storage |
HU02086 | S1 | An issue, in IP Quorum, may cause a Tier 2 recovery, during initial connection to a candidate device | 8.2.1.8 | IP Quorum |
HU02089 | S1 | Due to changes to quorum management, during an upgrade to v8.2.x, or later, there may be multiple warmstarts, with the possibility of a loss of access to data | 8.2.1.8 | System Update |
HU02121 | S1 | When the system changes from copyback to rebuild a failure to clear related metadata can cause multiple node warmstarts, with the possibility of a loss of access | 8.2.1.8 | Distributed RAID |
IT26257 | S1 | Starting a relationship, when the remote volume is offline, may result in a T2 recovery | 8.2.1.8 | HyperSwap |
IT30595 | S1 | A resource shortage in the RAID component can cause MDisks to be taken offline | 8.2.1.8 | RAID |
FLASH-29324 | S2 | Reading enclosure VPD can cause timeouts leading to node warmstarts | 8.2.1.8 | Reliability Availability Serviceability |
HU01836 | S2 | When an auxiliary volume is moved an issue with pausing the master volume can lead to node warmstarts | 8.2.1.8 | HyperSwap |
HU02049 | S2 | GUI session handling has an issue that can generate many exceptions, adversely impacting GUI performance | 8.2.1.8 | Graphical User Interface |
HU01880 | S3 | When a write, to a secondary volume, becomes stalled, a node at the primary site may warmstart | 8.2.1.8 | Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01936 | S3 | When shrinking a volume, that has host mappings, there may be recurring node warmstarts | 8.2.1.8 | Cache |
HU02021 | S3 | Disabling garbage collection may cause a node warmstart | 8.2.1.8 | Data Reduction Pools |
HU02085 | S3 | Freeze time of Global Mirror remote copy consistency groups may not be updated correctly in certain scenarios | 8.2.1.8 | Global Mirror |
IT30448 | S3 | If an IP Quorum app is killed, during the commit phase of a code upgrade, then that offline IP Quorum device cannot be removed, post upgrade | 8.2.1.8 | IP Quorum |
HU01998 | S1 | HIPER (Highly Pervasive): All SCSI command types can set volumes as busy resulting in I/O timeouts, and multiple node warmstarts, with the possibility of an offline I/O group. For more details refer to the following Flash | 8.2.1.6 | Hosts |
HU02014 | S1 | HIPER (Highly Pervasive): After a loss of power, where an AC3 node has a dead CMOS battery, it will fail to restart correctly. It is possible for both nodes in an I/O group to experience this issue | 8.2.1.6 | Reliability Availability Serviceability |
HU01888 & HU01997 | S1 | An issue with restore mappings, in the FlashCopy component, can cause an I/O group to warmstart | 8.2.1.6 | FlashCopy |
HU01933 | S1 | Under rare circumstances the Data Reduction Pool deduplication rehoming process can become truncated. Subsequent detection of inconsistent metadata can lead to offline Data Reduction Pools | 8.2.1.6 | Data Reduction Pools, Deduplication |
HU01985 | S1 | As a consequence of a Data Reduction Pool recovery bad metadata may be created. When the region of disk associated with the bad metadata is accessed there may be an I/O group warmstarts | 8.2.1.6 | Data Reduction Pools |
HU01989 | S1 | For large drives, bitmap scanning, during a rebuild, can timeout resulting in multiple node warmstarts, possibly leading to offline I/O groups | 8.2.1.6 | Distributed RAID |
HU02016 | S1 | A memory leak in the component that handles thin-provisioned MDisks can lead to an adverse performance impact with the possibility of offline MDisks | For more details refer to the following Flash8.2.1.6 | Backend Storage |
HU02027 | S1 | Fabric congestion can cause internal resource constraints, in 16Gb HBAs, leading to lease expiries | 8.2.1.6 | Reliability Availability Serviceability |
HU02043 | S1 | Collecting a snap can cause nodes to run out of boot drive space and go offline with node error 565 | 8.2.1.6 | Support Data Collection |
HU02045 | S1 | When a node is removed from the cluster, using CLI, it may still be shown as online in the GUI. If an attempt is made to shutdown this node, from the GUI, whilst it appears to be online, then the whole cluster will shutdown | 8.2.1.6 | Graphical User Interface |
HU01890 | S2 | FlashCopy mappings, from master volume to primary change volume, may become stalled when a T2 recovery occurs whilst the mappings are in a 'copying' state | 8.2.1.6 | Global Mirror with Change Volumes |
HU02037 | S2 | A FlashCopy consistency group, with a mix of mappings in different states, cannot be stopped | 8.2.1.6 | FlashCopy |
HU02046 | S2 | A timer issue within the support data collection function can cause truncated system report files for systems with AE2 enclosures | 8.2.1.6 | Support Data Collection |
HU02047 | S2 | When a system is upgraded to 8.2.1.5, a large number of events can be sent to call home | 8.2.1.6 | System Monitoring |
HU02055 | S2 | Creating a FlashCopy snapshot, in the GUI, does not set the same preferred node for both source and target volumes. This may adversely impact performance | 8.2.1.6 | FlashCopy |
HU02072 | S2 | An issue in the handling of email transmission can write a large file to the node boot drive. If this causes the boot drive to become full, the node will go offline with error 565 | 8.2.1.6 | System Monitoring |
FLASH-28947 | S3 | Active rebuild commands stuck after cross-bar failure | 8.2.1.6 | Reliability Availability Serviceability |
FLASH-29059 | S3 | Keep fans running in a canister that powers down to standby | 8.2.1.6 | Reliability Availability Serviceability |
FLASH-29060 | S3 | DMP support for keeping fans running in a canister that powers down to standby | 8.2.1.6 | GUI Fix Procedure |
FLASH-29061 | S3 | On system shutdown, power down fans | 8.2.1.6 | Reliability Availability Serviceability |
FLASH-29062 | S3 | Ramp fans to maximum speed when canister is missing | 8.2.1.6 | Reliability Availability Serviceability |
HU01843 | S3 | A node hardware issue can cause a CLI command to timeout resulting in a node warmstart | 8.2.1.6 | Command Line Interface |
HU01892 | S3 | LUNs of greater than 2TB, presented by HP XP7 storage controllers, are not supported | 8.2.1.6 | Backend Storage |
HU01974 | S3 | With all Remote Support Assistant connections closed, the GUI may show that a connection is still in progress | 8.2.1.6 | System Monitoring |
HU01978 | S3 | Unable to create HyperSwap volumes. The mkvolume command fails with CMMVC7050E error | 8.2.1.6 | HyperSwap |
HU01979 | S3 | The figure for used_virtualization, in the output of a lslicense command, may be unexpectedly large | 8.2.1.6 | Command Line Interface |
HU01982 | S3 | In an environment, with multiple IP Quorum servers, if the quorum component encounters a duplicate UID then a node may warmstart | 8.2.1.6 | IP Quorum |
HU01983 | S3 | Improve debug data capture to assist in determining the reason for a Data Reduction Pool to be taken offline | 8.2.1.6 | Data Reduction Pools |
HU01986 | S3 | An accounting issue in the FlashCopy component may cause node warmstarts | 8.2.1.6 | FlashCopy |
HU01991 | S3 | An issue in the handling of extent allocation, in the Data Reduction Pool component, can cause a node warmstart | 8.2.1.6 | Data Reduction Pools |
HU02029 | S3 | An issue with the SSMTP process may result in failed callhome, inventory reporting and user notifications. A testemail command will fail with a CMMVC9051E error | 8.2.1.6 | System Monitoring |
HU02039 | S3 | An issue in the management steps of Data Reduction Pool recovery may lead to a node warmstart | 8.2.1.6 | Data Reduction Pools |
HU02067 | S3 | If multiple recipients are specified, for callhome emails, then no callhome emails will be sent | 8.2.1.6 | System Monitoring |
HU02129 | S3 | GUI drive filtering fails with "An error occurred loading table data" | 8.2.1.6 | Graphical User Interface |
HU02007 | S1 | HIPER (Highly Pervasive): During volume migration an issue in the handling of old to new extents transfer can lead to cluster-wide warmstarts | 8.2.1.5 | Storage Virtualisation |
HU02009 | S1 | Systems which are using Data Reduction Pools, with the maximum possible extent size of 8GB, and which experience a very specific I/O workload, may experience an issue due to garbage collection. This can cause repeated node warmstarts and loss of access to data | 8.2.1.5 | Data Reduction Pools |
HU02011 | S2 | When a node warmstart occurs on a system using Data Reduction Pools, there is a small possibility that the node will not automatically return online. If the partner node is also offline, this can cause temporary loss of access to data | 8.2.1.5 | Data Reduction Pools |
HU02012 | S2 | Under certain I/O workloads the garbage collection process can adversely impact volume write response times | 8.2.1.5 | Data Reduction Pools |
HU01865 | S1 | HIPER (Highly Pervasive): When creating a HyperSwap relationship, using addvolumecopy (or similar methods), the system should perform a synchronisation operation to copy the data of the original copy to the new copy. In some cases this synchronisation is skipped, leaving the new copy with bad data (all zeros) | 8.2.1.4 | HyperSwap |
HU01918 | S1 | HIPER (Highly Pervasive): Where Data Reduction Pools have been created on earlier code levels, upgrading the system, to an affected release, can cause an increase in the level of concurrent flushing to disk. This may result in a loss of access to data. For more details refer to the following Flash | 8.2.1.4 | Data Reduction Pools |
HU02008 | S1 | HIPER (Highly Pervasive): When a DRAID rebuild occurs, occasionally a RAID deadlock condition can be triggered by a particular type of I/O workload. This can lead to repeated node warmstarts and a loss of access to data | 8.2.1.4 | Distributed RAID |
HU01887 | S1 | In circumstances where host configuration data becomes inconsistent, across nodes, an issue in the CLI policing code may cause multiple warmstarts | 8.2.1.4 | Command Line Interface, Host Cluster |
HU01900 | S1 | Executing a command, that can result in a shared mapping being created or destroyed, for an individual host, in a host cluster, without that command applying to all hosts in the host cluster, may lead to multiple node warmstarts with the possibility of a T2 recovery | 8.2.1.4 | Host Cluster |
HU01910 | S1 | When FlashCopy mappings are created, with a grain size of 64KB, it is possible for an overflow condition in the bitmap to occur. This can resulting in multiple node warmstarts with a possible loss of access to data | 8.2.1.4 | FlashCopy |
HU01928 | S1 | When two IOs attempt to access the same address, the state of the data may be incorrectly set to invalid causing offline volumes and, possibly, offline pools | 8.2.1.4 | Data Reduction Pools |
HU01987 | S1 | During SAN fabric power maintenance an AC3 cluster may breech resource limits, on the remaining node to node links, resulting in system-wide lease expiry | 8.2.1.4 | Reliability Availability Serviceability |
HU02000 | S1 | Data Reduction Pools may go offline due to a timing issue in metadata handling | 8.2.1.4 | Data Reduction Pools |
HU02013 | S1 | A race condition, between the extent invalidation and destruction, in the garbage collection process, may cause a node warmstart with the possibility of offline volumes | 8.2.1.4 | Data Reduction Pools |
HU02025 | S1 | An issue with metadata handling, where a pool has been taken offline, may lead to an out of space condition, in that pool, preventing its return to operation | 8.2.1.4 | Data Reduction Pools |
HU01886 | S2 | The Unmap function can leave volume extents, that have not been freed, preventing managed disk and pool removal | 8.2.1.4 | SCSI Unmap |
HU01941 | S2 | After upgrading the system to v8.2, or later, when expanding a mirrored volume, the format of additional space may become stalled | 8.2.1.4 | Volume Mirroring |
HU01944 | S2 | Proactive host failover not waiting for 25 seconds before allowing nodes to go offline during upgrades or maintenance | 8.2.1.4 | Reliability Availability Serviceability |
HU01972 | S2 | When an array is in a quiescing state, for example where a member has been deleted, I/O may become pended leading to multiple warmstarts | 8.2.1.4 | RAID, Distributed RAID |
HU00744 | S3 | Single node warmstart due to an accounting issue within the cache component | 8.2.1.4 | Cache |
HU01485 | S3 | When a AC3 node is started, with only one PSU powered, powering up the other PSU will not extinguish the Power Fault LED. Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed |
8.2.1.4 | System Monitoring |
HU01659 | S3 | On AC3 systems the Node Fault LED may be seen to flash in the absence of an error condition. Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed |
8.2.1.4 | System Monitoring |
HU01857 | S3 | Improved validation of user input in GUI | 8.2.1.4 | Graphical User Interface |
HU01860 | S3 | During garbage collection the flushing of extents may become stuck leading to a timeout and a single node warmstart | 8.2.1.4 | Data Reduction Pools |
HU01869 | S3 | Volume copy deletion in a Data Reduction Pool triggered by rmvdiskcopy rmvolumecopy or addvdiskcopy -autodelete (or similar) may become stalled with the copy being left in "deleting" status | 8.2.1.4 | Data Reduction Pools |
HU01911 | S3 | The "System Overview" screen, in the GUI, may display nodes in the wrong site | 8.2.1.4 | Graphical User Interface |
HU01912 | S3 | Systems with iSCSI-attached controllers may see node warmstarts due to I/O request timeouts | 8.2.1.4 | Backend Storage |
HU01915 & IT28654 | S3 | For systems, with encryption enabled, that are using key servers to manage encryption keys, the system may fail to connect to the key servers if the server's SSL certificates is part of a chain of trust | 8.2.1.4 | Encryption |
HU01916 | S3 | The GUI Dashboard and the CLI lssystem command report physical capacity incorrectly | 8.2.1.4 | Graphical User Interface, Command Line Interface |
HU01926 | S3 | When a node with 32GB of RAM is upgraded to v8.2.1 it may experience a warmstart resulting in a failed upgrade | 8.2.1.4 | System Update |
HU01959 | S3 | An timing window issue in the Thin Provisioning component can cause a node warmstart | 8.2.1.4 | FlashCopy, Thin Provisioning |
HU01962 | S3 | When Call Home servers return an invalid message it can be incorrectly reported as an error 3201 in the Event Log | 8.2.1.4 | System Monitoring |
HU01976 | S3 | A new MDisk array may not be encrypted even though encryption is enabled on the system | 8.2.1.4 | Encryption |
HU02001 | S3 | During a system upgrade an issue in callhome may cause a node warmstart, stalling the upgrade | 8.2.1.4 | System Monitoring |
HU02002 | S3 | On busy systems, diagnostic data collection may not complete correctly producing livedumps with missing pages | 8.2.1.4 | Support Data Collection |
HU02019 | S3 | When the master and auxiliary volumes, in a relationship, have the same name it is not possible, in the GUI, to determine which is master or auxiliary | 8.2.1.4 | Graphical User Interface |
HU02166 | S3 | A timing window issue, in RAID code that handles recovery after a drive has been taken out of sync, due to a slow I/O, can cause a single node warmstart | 8.2.1.4 | RAID |
IT28433 | S3 | Timing window issue in the Data Reduction Pool rehoming component can cause a single node warmstart | 8.2.1.4 | Data Reduction Pools |
IT28728 | S3 | Email alerts will not work where the mail server does not allow unqualified client host names | 8.2.1.4 | System Monitoring |
HU01932 | S1 | When a rmvdisk command initiates a Data Reduction Pool rehoming process, any I/O to deduplicated data, referenced to data on the removed volume, may cause multiple warmstarts leading to a loss of access. | 8.2.1.2 | Deduplication |
FLASH-28023 | S2 | V9000 no longer sending heartbeats. | 8.2.1.2 | Reliability Availability Serviceability |
HU01920 | S1 | An issue in the garbage collection process can cause node warmstarts and offline pools | 8.2.1.1 | Data Reduction Pools |
HU01492 & HU02024 | S1 | HIPER (Highly Pervasive): All ports of a 16Gb HBA can be affected when a single port is congested. This can lead to lease expiries if all ports used for inter-node communication are on the same FC adapter | 8.2.1.0 | Reliability Availability Serviceability |
HU01617 | S1 | HIPER (Highly Pervasive): Due to a timing window issue, stopping a FlashCopy mapping, with the -autodelete option, may result in a Tier 2 recovery | 8.2.1.0 | FlashCopy |
HU01873 | S1 | HIPER (Highly Pervasive): Deleting a volume, in a Data Reduction Pool, while volume protection is enabled and when the volume was not explicitly unmapped, before deletion, may result in simultaneous node warmstarts. For more details refer to the following Flash | 8.2.1.0 | Data Reduction Pools |
HU01743 | S1 | Where hosts are directly attached, a mishandling of the login process, by the fabric controller, may result in dual node warmstarts | 8.2.1.0 | Hosts |
HU01758 | S1 | After an unexpected power loss, all nodes, in a cluster, may warmstart repeatedly, necessitating a Tier 3 recovery | 8.2.1.0 | RAID |
HU01799 | S1 | Timing window issue can affect operation of the HyperSwap addvolumecopy command causing all nodes to warmstart | 8.2.1.0 | HyperSwap |
HU01825 | S1 | Invoking a chrcrelationship command when one of the relationships, in a consistency group, is running in the opposite direction to the others may cause a node warmstart followed by a Tier 2 recovery | 8.2.1.0 | FlashCopy |
HU01828 | S1 | Node warmstarts may occur during deletion of deduplicated volumes due to a timing-related issue | 8.2.1.0 | Deduplication |
HU01833 | S1 | If both nodes, in an I/O group, start up together a timing window issue may occur, that would prevent them running garbage collection, leading to a related Data Reduction Pool running out of space | 8.2.1.0 | Data Reduction Pools |
HU01845 | S1 | If the execution of a rmvdisk -force command, for the FlashCopy target volume in a GMCV relationship, coincides with the start of a GMCV cycle all nodes may warmstart | 8.2.1.0 | Global Mirror with Change Volumes |
HU01847 | S1 | FlashCopy handling of medium errors, across a number of drives on backend controllers, may lead to multiple node warmstarts | 8.2.1.0 | FlashCopy |
HU01848 | S1 | During an upgrade, systems with a large AIX VIOS setup may have multiple node warmstarts with the possibility of a loss of access to data | 8.2.1.0 | System Update |
HU01850 | S1 | When the last deduplication-enabled volume copy, in a Data Reduction Pool, is deleted, the pool may go offline temporarily | 8.2.1.0 | Data Reduction Pools, Deduplication |
HU01851 | S1 | When a deduplicated volume is deleted there may be multiple node warmstarts and offline pools | 8.2.1.0 | Deduplication |
HU01855 | S1 | Clusters using Data Reduction Pools can experience multiple warmstarts on all nodes putting them in a service state | 8.2.1.0 | Data Reduction Pools |
HU01862 | S1 | When a Data Reduction Pool is removed, and the -force option is specified, there may be a temporary loss of access | 8.2.1.0 | Data Reduction Pools |
HU01876 | S1 | Where systems are connected to controllers, that have FC ports that are capable of acting as initiators and targets, when NPIV is enabled then node warmstarts can occur | 8.2.1.0 | Backend Storage |
HU01878 | S1 | During an upgrade, from v7.8.1 or earlier to v8.1.3 or later, if an MDisk goes offline then, at completion, all volumes may go offline | 8.2.1.0 | System Update |
HU01885 | S1 | As writes are made to a Data Reduction Pool it is necessary to allocate new physical capacity. Under unusual circumstances it is possible for the handling of an expansion request to stall further I/O leading to node warmstarts | 8.2.1.0 | Data Reduction Pools |
HU01913 | S1 | A timing window issue in the DRAID6 rebuild process can cause node warmstarts with the possibility of a loss of access | 8.2.1.0 | Distributed RAID |
HU01957 | S1 | Due to an issue in Data Reduction Pools, when the system attempts an upgrade, there may be node warmstarts | 8.2.1.0 | Data Reduction Pools, System Update |
HU01965 | S1 | A timing window issue in the deduplication component can lead to I/O timeouts, and a node warmstart, with the possibility of an offline mdisk group | 8.2.1.0 | Deduplication |
HU02042 | S1 | An issue in the handling of metadata, after a Data Reduction Pool recovery operation, can lead to repeated node warmstarts, putting an I/O group into a service state | 8.2.1.0 | Data Reduction Pools |
IT25850 | S1 | I/O performance may be adversely affected towards the end of DRAID rebuilds. For some systems there may be multiple warmstarts leading to a loss of access | 8.2.1.0 | Distributed RAID |
IT27460 | S1 | Lease expiry can occur between local nodes when remote connection is lost, due to the mishandling of messaging credits | 8.2.1.0 | Reliability Availability Serviceability |
IT29040 | S1 | Occasionally a DRAID rebuild, with drives of 8TB or more, can encounter an issue which causes node warmstarts and potential loss of access | 8.2.1.0 | RAID, Distributed RAID |
FLASH-27506 | S2 | Improved RAID error handling for unresponsive flash modules to prevent rare data error | 8.2.1.0 | RAID |
HU01507 | S2 | Until the initial synchronisation process completes, high system latency may be experienced when a volume is created with two compressed copies or when space-efficient copy is added to a volume with an existing compressed copy | 8.2.1.0 | Volume Mirroring |
HU01661 | S2 | A cache-protection mechanism flag setting can become stuck, leading to repeated stops of consistency group synching | 8.2.1.0 | HyperSwap |
HU01733 | S2 | Canister information, for the High Density Expansion Enclosure, may be incorrectly reported | 8.2.1.0 | Reliability Availability Serviceability |
HU01761 | S2 | Entering multiple addmdisk commands, in rapid succession, to more than one storage pool, may cause node warmstarts | 8.2.1.0 | Backend Storage |
HU01797 | S2 | Hitachi G1500 backend controllers may exhibit higher than expected latency | 8.2.1.0 | Backend Storage |
HU01810 | S2 | Deleting volumes, or using FlashCopy/Global Mirror with Change Volumes, in a Data Reduction Pool, may impact the performance of other volumes in the pool. | 8.2.1.0 | Data Reduction Pools |
HU01837 | S2 | In systems, where a VVols metadata volume has been created, an upgrade to v8.1.3 or later will cause a node warmstart, stalling the upgrade | 8.2.1.0 | VVols |
HU01839 | S2 | Where a VMware host is being served volumes, from two different controllers, and an issue, on one controller, causes the related volumes to be taken offline then I/O performance, for the volumes from the other controller, will be adversely affected | 8.2.1.0 | Hosts |
HU01842 | S2 | Bursts of I/O to Samsung Read-Intensive Drives can be interpreted as dropped frames, against the resident slots, leading to redundant drives being incorrectly failed | 8.2.1.0 | Drives |
HU01846 | S2 | Silent battery discharge condition will, unexpectedly, take a SVC node offline, putting it into a 572 service state | 8.2.1.0 | Reliability Availability Serviceability |
HU01852 | S2 | The garbage collection rate can lead to Data Reduction Pools running out of space even though reclaimable capacity is available | 8.2.1.0 | Data Reduction Pools |
HU01858 | S2 | Total used capacity of a Data Reduction Pool within a single I/O group is limited to 256TB. Garbage collection does not correctly recognise this limit. This may lead to a pool running out of free capacity and going offline | 8.2.1.0 | Data Reduction Pools |
HU01881 | S2 | An issue within the compression card in FS9100 systems can result in the card being incorrectly flagged as failed leading to warmstarts | 8.2.1.0 | Compression |
HU01883 | S2 | Config node processes may consume all available memory, leading to node warmstarts. This can be caused, for example, by large numbers of concurrent SSH connections being opened | 8.2.1.0 | Reliability Availability Serviceability |
HU01907 | S2 | An issue in the handling of the power cable sense registers can cause a node to be put into service state with a 560 error | 8.2.1.0 | Reliability Availability Serviceability |
HU02076 | S2 | For systems with AE2 enclosures, error 1007 (Fewer bus ports operational) should generate a callhome | 8.2.1.0 | System Monitoring |
HU00921 | S3 | A node warmstart may occur when an MDisk state change gives rise to duplicate discovery processes | 8.2.1.0 | |
HU01276 | S3 | An issue in the handling of debug data from the FC adapter can cause a node warmstart | 8.2.1.0 | Reliability Availability Serviceability |
HU01523 | S3 | An issue with FC adapter initialisation can lead to a node warmstart | 8.2.1.0 | Reliability Availability Serviceability |
HU01564 | S3 | FlashCopy maps cleaning process is not monitoring the grains correctly which may cause FlashCopy maps to not stop | 8.2.1.0 | FlashCopy |
HU01571 | S3 | An upgrade can become stalled due to a node warmstart | 8.2.1.0 | System Update |
HU01657 | S3 | The 16Gb FC HBA firmware may experience an issue, with the detection of unresponsive links, leading to a single node warmstart | 8.2.1.0 | Reliability Availability Serviceability |
HU01667 | S3 | A timing-window issue, in the remote copy component, may cause a node warmstart | 8.2.1.0 | Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01719 | S3 | Node warmstart due to a parity error in the HBA driver firmware | 8.2.1.0 | Reliability Availability Serviceability |
HU01737 | S3 | On the "Update System" screen, for "Test Only", if a valid code image is selected, in the "Run Update Test Utility" dialog, then clicking the "Test" button will initiate a system update | 8.2.1.0 | System Update |
HU01751 | S3 | When RAID attempts to flag a strip as bad, and that strip has already been flagged, a node may warmstart | 8.2.1.0 | RAID |
HU01760 | S3 | FlashCopy map progress appears to be stuck at zero percent | 8.2.1.0 | FlashCopy |
HU01765 | S3 | Node warmstart may occur when there is a delay to I/O at the secondary site | 8.2.1.0 | Global Mirror, Global Mirror with Change Volumes, Metro Mirror |
HU01772 | S3 | The mail queue may become blocked preventing the transmission of event log messages | 8.2.1.0 | System Monitoring |
HU01784 | S3 | If a cluster using IP quorum experiences a site outage, the IP quorum device may become invalid. Restarting the quorum application will resolve the issue | 8.2.1.0 | HyperSwap, Quorum |
HU01786 | S3 | An issue in the monitoring of SSD write endurance can result in false 1215/2560 errors in the Event Log | 8.2.1.0 | Drives |
HU01791 | S3 | Using the chhost command will remove stored CHAP secrets | 8.2.1.0 | iSCSI |
HU01807 | S3 | The lsfabric command may show incorrect local node id and local node name for some Fibre Channel logins | 8.2.1.0 | Command Line Interface |
HU01811 | S3 | DRAID rebuilds, for large (>10TB) drives, may require lengthy metadata processing leading to a node warmstart | 8.2.1.0 | Distributed RAID |
HU01815 | S3 | In Data Reduction Pools, volume size is limited to 96TB | 8.2.1.0 | Data Reduction Pools |
HU01817 | S3 | Volumes used for VVols metadata or cloud backup, that are associated with a FlashCopy mapping, cannot be included in any further FlashCopy mappings | 8.2.1.0 | FlashCopy |
HU01821 | S3 | An attempt to upgrade a two-node enhanced stretched cluster fails due to incorrect volume dependencies | 8.2.1.0 | System Update, Data Reduction Pools |
HU01832 | S3 | Creation and distribution of the config file may cause an out-of-memory condition, leading to a node warmstart | 8.2.1.0 | Reliability Availability Serviceability |
HU01849 | S3 | An excessive number of SSH sessions may lead to a node warmstart | 8.2.1.0 | System Monitoring |
HU01856 | S3 | A garbage collection process can time out waiting for an event in the partner node resulting in a node warmstart | 8.2.1.0 | Data Reduction Pools |
HU01863 | S3 | In rare circumstances, a drive replacement may result in a "ghost drive" (i.e. a drive with the same ID as the replaced drive stuck in a permanently offline state) | 8.2.1.0 | Drives |
HU01871 | S3 | An issue with bitmap synchronisation can lead to a node warmstart | 8.2.1.0 | Data Reduction Pools |
HU01879 | S3 | Latency induced by DWDM inter-site links may result in a node warmstart | 8.2.1.0 | |
HU01893 | S3 | Excessive reporting frequency of NVMe drive diagnostics generates large numbers of callhome events | 8.2.1.0 | Drives |
HU01895 | S3 | Where a banner has been created, without a new line at the end, any subsequent T4 recovery will fail | 8.2.1.0 | Distributed RAID |
HU01981 | S3 | Although an issue in the 16Gb HBA firmware is handled correctly it can still cause a node warmstart | 8.2.1.0 | Reliability Availability Serviceability |
HU02028 | S3 | An issue, with timer cancellation, in the Remote Copy component may cause a node warmstart | 8.2.1.0 | Metro Mirror, Global Mirror, Global Mirror with Change Volumes |
IT19561 | S3 | An issue with register clearance in the FC driver code may cause a node warmstart | 8.2.1.0 | Reliability Availability Serviceability |
IT25457 | S3 | Attempting to remove a copy of a volume, which has at least one image mode copy and at least one thin/compressed copy, in a Data Reduction Pool will always fail with a CMMVC8971E error | 8.2.1.0 | Data Reduction Pools |
IT25970 | S3 | After a FlashCopy consistency group is started a node may warmstart | 8.2.1.0 | FlashCopy |
IT26049 | S3 | An issue with CPU scheduling may cause the GUI to respond slowly | 8.2.1.0 | Graphical User Interface |
Please refer to the Concurrent Compatibility and Code Cross Reference for Spectrum Virtualize page for guidance when planning a system upgrade.
Description | Link |
---|---|
Support Website | IBM Knowledge Center |
IBM FlashSystem Fix Central | V9000 |
Updating the system | IBM Knowledge Center |
IBM Redbooks | Redbooks |
Contacts | IBM Planetwide |