SVAPAR-130438 |
All |
HIPER
|
Upgrading a system to 8.6.2 or higher with a single portset assigned to an IP replication partnership may cause all nodes to warmstart when making a change to the partnership.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems with IP replication partnerships. |
Trigger |
Upgrading a system with a single portset assigned to an IP replication partnership and making a change to the partnership. |
Workaround |
None |
|
8.6.0.5 |
IP Replication |
SVAPAR-115129 |
All |
Critical
|
A node can warmstart when its I/O group partner node is removed due to an internal software counter discrepancy. This can lead to temporary loss of access.
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems configured with Data Reduction Pools. |
Trigger |
One node of an I/O group pair is removed, for example during an upgrade |
Workaround |
None |
|
8.6.0.5 |
Data Reduction Pools |
SVAPAR-117005 |
All |
Critical
|
A system may run an automatic cluster recovery, and warmstart all nodes, if Policy-based Replication is disabled on the partnership before removing the replication policy.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using Policy-based Replication |
Trigger |
Policy-based Replication is disabled on the partnership before removing the replication policy. |
Workaround |
Remove the Replication policy before disabling Policy-based Replication on the partnership. |
|
8.6.0.5 |
Policy-based Replication |
SVAPAR-120397 |
All |
Critical
|
A node may not shutdown cleanly on loss of power if it contains 25Gb Ethernet adapters, necessitating a system recovery.
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with 25Gb Ethernet adapters. |
Trigger |
Loss of power to the system. |
Workaround |
None |
|
8.6.0.5 |
Reliability Availability Serviceability |
SVAPAR-128912 |
All |
Critical
|
A T2 recovery may occur when attempting to take a snapshot from a volume group that contains volumes from multiple I/O groups, and one of the I/O groups is offline.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems with volume groups containing volumes from multiple I/O groups |
Trigger |
Taking a snapshot while one I/O group is offline. This includes snapshots taken automatically based on a scheduled policy |
Workaround |
Suspend the snapshot policy |
|
8.6.0.5 |
FlashCopy, Safeguarded Copy & Safeguarded Snapshots |
SVAPAR-130553 |
All |
Critical
|
Converting a 3-Site AuxFar volume to HyperSwap results in multiple node asserts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
3-Site system with HyperSwap in the AuxFar site configured |
Trigger |
Converting a 3-Site AuxFar volume to HyperSwap |
Workaround |
Avoid converting 3-Site AuxFar volumes to HyperSwap |
|
8.6.0.5 |
3-Site using HyperSwap or Metro Mirror, HyperSwap |
SVAPAR-131228 |
All |
Critical
|
A RAID array temporarily goes offline due to delays in fetching the encryption key when a node starts up.
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using encryption configured to use encryption key servers |
Trigger |
Start up of a node |
Workaround |
Reduce the number of configured key servers to two. |
|
8.6.0.5 |
Distributed RAID, Encryption, RAID |
SVAPAR-131259 |
All |
Critical
|
Removal of the replication policy after the volume group was set to be independent exposed an issue that resulted in the FlashCopy internal state becoming incorrect, this meant subsequent FlashCopy actions failed incorrectly.
(show details)
Symptom |
None |
Environment |
Systems configured with Volume Groups, Policy-based replication and FlashCopy. |
Trigger |
A volume group used for Policy-based Replication has its target side made independent. Then the volume group has its replication policy removed. After these actions, if any volumes in the volume group are used for FlashCopy, it will hit the error. |
Workaround |
None |
|
8.6.0.5 |
FlashCopy, Policy-based Replication |
SVAPAR-131648 |
All |
Critical
|
Multiple node warmstarts may occur when starting an incremental FlashCopy map that uses a replication target volume as its source, and the change volume is used to keep a consistent image.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems with Policy-Based Replication and incremental FlashCopy maps on the same volumes |
Trigger |
Starting an incremental FlashCopy mapping while the change volume is used to keep a consistent image |
Workaround |
Avoid starting incremental maps whilst the change volume is used to keep a consistent image |
|
8.6.0.5 |
FlashCopy, Policy-based Replication |
SVAPAR-141094 |
All |
Critical
|
On power failure, FS50xx systems with a 25Gb ROCE adapters may fail to gracefully shutdown, causing loss of cache data.
(show details)
Symptom |
Loss of Access to Data |
Environment |
FS50xx systems with a 25Gb ROCE adapters |
Trigger |
Power failure |
Workaround |
None |
|
8.6.0.5 |
Reliability Availability Serviceability |
HU02159 |
All |
High Importance
|
A rare issue caused by unexpected I/O in the upper cache can cause a node to warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Offline target disk |
Trigger |
Target disk must be offline for this situation to occur. |
Workaround |
None |
|
8.6.0.5 |
Cache |
SVAPAR-111173 |
All |
High Importance
|
Loss of access when two drives experience slowness at the same time
(show details)
Symptom |
Loss of Access to Data |
Environment |
Can occur on any SAS based system, including SAS expansion enclosures. |
Trigger |
Heavily loaded SAS spinning drives |
Workaround |
Try and reduce the work load |
|
8.6.0.5 |
RAID |
SVAPAR-117457 |
All |
High Importance
|
A hung condition in Remote Receive IOs (RRI) for volume groups can lead to warmstarts on multiple nodes.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any system that uses Policy-based Replication |
Trigger |
None |
Workaround |
None |
|
8.6.0.5 |
Policy-based Replication |
SVAPAR-119799 |
FS9500, SVC |
High Importance
|
Inter-node resource queuing on SV3 I/O groups, causes high write response time.
(show details)
Symptom |
Performance |
Environment |
This can occur on environments that use clustering over Fibre Channel, or the SAN is not optimal, or have geographically dispersed sites for Enhanced Stretched cluster or Hyperswap |
Trigger |
High intra-cluster round trip time. |
Workaround |
None |
|
8.6.0.5 |
Performance |
SVAPAR-120630 |
All |
High Importance
|
An MDisk may go offline due to I/O timeouts caused by an imbalanced workload distribution towards the resources in DRP, whilst FlashCopy is running at a high copy rate within DRP, and the target volume is dedup.
(show details)
Symptom |
Offline Volumes |
Environment |
Any system running FlashCopy, with a deduplicated target volume in DRP. |
Trigger |
None |
Workaround |
None |
|
8.6.0.5 |
Data Reduction Pools |
SVAPAR-127845 |
All |
High Importance
|
Attempting to create a second I/O group, in the two `Caching I/O Group` dropdowns on the `Define Volume Properties` modal of `Create Volumes` results in error `CMMVC8709E the iogroups of cache memory storage are not in the same site as the storage groups`.
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
Attempting to create a second I/O group, in the two `Caching I/O Group` dropdowns on the `Define Volume Properties` modal of `Create Volumes` |
Workaround |
None |
|
8.6.0.5 |
GUI Fix Procedure, Graphical User Interface |
SVAPAR-127869 |
All |
High Importance
|
Multiple node warmstarts may occur, due to a rarely seen timing window, when quorum disk I/O is submitted but there is no backend mdisk Logical Unit association that has been discovered by the agent for that quorum disk.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
None |
Trigger |
None |
Workaround |
None |
|
8.6.0.5 |
Quorum |
SVAPAR-128914 |
All |
High Importance
|
A CMMVC9859E error will occur when trying to use 'addvolumecopy' to create Hyperswap volume from a VDisk with existing snapshots
(show details)
Symptom |
Configuration |
Environment |
Systems configured with Hyperswap topology with snapshots |
Trigger |
Trying to create a Hyperswap volume from an existing VDisk |
Workaround |
None |
|
8.6.0.5 |
HyperSwap |
SVAPAR-129318 |
All |
High Importance
|
A Storage Virtualize cluster configured without I/O group 0 is unable to send performance metrics
(show details)
Symptom |
Configuration |
Environment |
Any cluster that does not have I/O group 0 configured |
Trigger |
None |
Workaround |
Configure I/O group 0 |
|
8.6.0.5 |
Performance |
SVAPAR-131651 |
All |
High Importance
|
Policy-based Replication got stuck after both nodes in the I/O group on a target system restarted at the same time
(show details)
Symptom |
Loss of Redundancy |
Environment |
Target system using Policy-based Replication. |
Trigger |
Both nodes in the I/O group on a target system restart at the same time. |
Workaround |
None |
|
8.6.0.5 |
Policy-based Replication |
SVAPAR-136427 |
All |
High Importance
|
When deleting multiple older snapshots versions, whilst simultaneously creating new snapshots, the system can run out of bitmap space resulting in a bad snapshot map, repeated asserts, and a loss of access.
(show details)
Symptom |
Loss of Access to Data |
Environment |
This can affect any system configured with FlashCopy |
Trigger |
Deletion of multiple mid-age snapshots and creating new snapshots. |
Workaround |
Avoid deleting multiple older snapshots versions, whilst simultaneously creating new snapshots |
|
8.6.0.5 |
FlashCopy |
SVAPAR-137265 |
All |
High Importance
|
Error when attempting to delete a HyperSwap volume with snapshots
(show details)
Symptom |
Configuration |
Environment |
Systems configured to use Hyperswap |
Trigger |
None |
Workaround |
None |
|
8.6.0.5 |
FlashCopy |
SVAPAR-141996 |
All |
High Importance
|
Policy-based replication may not perform the necessary background synchronization to maintain an up to date copy of data at the DR site.
(show details)
Symptom |
Loss of Redundancy |
Environment |
Volumegroups configured for Policy-based Replication may need to perform background synchronisation of data. This process is typically driven by the nodes in the recovery system. |
Trigger |
None |
Workaround |
None |
|
8.6.0.5 |
Policy-based Replication |
SVAPAR-89331 |
All |
High Importance
|
Systems running 8.5.2 or higher using IP replication with compression may have low replication bandwidth and high latency due to an issue with the way the data is compressed.
(show details)
Symptom |
Performance |
Environment |
Systems using IP replication compression |
Trigger |
None |
Workaround |
Disable compression on IP partnership |
|
8.6.0.5 |
IP Replication |
SVAPAR-111991 |
All |
Suggested
|
Attempting to create a truststore fails with a CMMVC5711E error if the certificate file does not end with a newline character
(show details)
Symptom |
Configuration |
Environment |
Systems using policy-based Replication, secured IP partnerships or VASA. |
Trigger |
Attempting to create a truststore |
Workaround |
Ensure the certificate file ends with a newline character |
|
8.6.0.5 |
IP Replication, Policy-based Replication, vVols |
SVAPAR-114145 |
All |
Suggested
|
A timing issue triggered by disabling an IP partnership's compression state while replication is running may cause a node to warmstart.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems configured with IP replication. |
Trigger |
Disabling an IP partnership's compression state whilst replication is running. |
Workaround |
Stop replication before changing the IP partnership's compression state |
|
8.6.0.5 |
IP Replication |
SVAPAR-127835 |
All |
Suggested
|
A node may warmstart due to invalid RDMA receive size of zero.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems configure with NVMeF over RDMA |
Trigger |
RDMA receiving size of zero detected due to API error |
Workaround |
None |
|
8.6.0.5 |
NVMe |
SVAPAR-129274 |
All |
Suggested
|
When running the 'mkvolumegroup' command, a warmstart of the Config node may occur.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems configured with volume group snapshots |
Trigger |
Running the 'mkvolumegroup' command with the options '-fromsourcegroup', '-snapshot' and '-type clone'. |
Workaround |
Review whether there is sufficient free FlashCopy bitmap space before running the 'mkvolumegroup' command or performing the GUI action. Increase if necessary. |
|
8.6.0.5 |
FlashCopy, Thin Provisioning |
SVAPAR-131212 |
All |
Suggested
|
The GUI partnership properties dialog crashes if the issuer certificate does not have an organization field
(show details)
Symptom |
Configuration |
Environment |
Systems with a partnership |
Trigger |
Opening the partnership property dialog or creating a partnership while the issuer certificate has no organization field |
Workaround |
When using an externally signed certificate, make sure the issuer certificate has a non-empty organization (O=) field |
|
8.6.0.5 |
Policy-based Replication |
SVAPAR-131807 |
All |
Suggested
|
The orchestrator for Policy-Based Replication is not running, preventing replication from being configured. Attempting to configure replication may cause a single node warmstart.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with MTM 4666-AG8 |
Trigger |
Setting up Policy-Based Replication |
Workaround |
None |
|
8.6.0.5 |
Policy-based Replication |
SVAPAR-131865 |
All |
Suggested
|
A system may encounter communication issues when being configured with IPv6.
(show details)
Symptom |
Configuration |
Environment |
Systems using IPV6 |
Trigger |
None |
Workaround |
None |
|
8.6.0.5 |
|
SVAPAR-131994 |
All |
Suggested
|
When implementing Safeguarded Copy, the associated child pool may run out of space, which can cause multiple Safeguarded Copies to go offline. This can cause the node to warmstart.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems that are being configured with Safeguarded Copy |
Trigger |
Not enough space dedicated to FlashCopy target |
Workaround |
Configure more space for FlashCopy |
|
8.6.0.5 |
Safeguarded Copy & Safeguarded Snapshots |
SVAPAR-132011 |
All |
Suggested
|
In rare situations, a hosts WWPN may show incorrectly as still logged into the storage even though it is not. This can cause the host to incorrectly appear as degraded.
(show details)
Symptom |
Configuration |
Environment |
Systems using Fibre Channel host connections. |
Trigger |
Disabling or removing a host fibre channel connection. |
Workaround |
None |
|
8.6.0.5 |
Fibre Channel, Reliability Availability Serviceability |
SVAPAR-132072 |
All |
Suggested
|
A node may assert due to a Fibre Channel port constantly flapping between the FlashSystem and the host.
(show details)
Symptom |
Single Node Warmstart |
Environment |
None |
Trigger |
Fibre Channel port constantly flapping between the FlashSystem and the host. |
Workaround |
Replace either the SFP or Fiber optic, or both, between FlashSystem and the host |
|
8.6.0.5 |
Fibre Channel |
SVAPAR-135000 |
All |
Suggested
|
A low-probability timing window in memory management code may cause a single-node warmstart at upgrade completion.
(show details)
Symptom |
Single Node Warmstart |
Environment |
None |
Trigger |
Software upgrade |
Workaround |
None |
|
8.6.0.5 |
System Update |
SVAPAR-135742 |
All |
Suggested
|
A temporary network issue may cause unexpected 1585 DNS connection errors after upgrading to 8.6.0.4, 8.6.3.0 or 8.7.0.0. This is due to a shorter DNS request timeout in these PTFs.
(show details)
Symptom |
Configuration |
Environment |
Systems with a DNS server configured. |
Trigger |
Temporary network issue causing a DNS request timeout. |
Workaround |
Check the network for any issues that may cause packet loss, which could lead to a timeout on DNS requests. 'traceroute' command from the cluster to the DNS server can help determine if there are routes that are slower than others. |
|
8.6.0.5 |
Reliability Availability Serviceability |
SVAPAR-136172 |
All |
Suggested
|
VMware vCentre reports a disk expansion failure, prior to changing the provisioning policy.
(show details)
Symptom |
None |
Environment |
Systems that have VMware virtual volumes |
Trigger |
Expanding virtual volume within VMWare vCenter |
Workaround |
Ensure pool provisioning permits volume expansion |
|
8.6.0.5 |
vVols |
SVAPAR-137241 |
All |
Suggested
|
When attempting to create a Hyperswap volume via the GUI, when the preferred site is in the secondary data centre, a CMMVC8709E 'the iogroups of cache memory storage are not in the same site as the storage groups' failure occurs.
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
Creating a Hyperswap volume from the GUI, when the preferred site is in the secondary data centre |
Workaround |
None |
|
8.6.0.5 |
GUI Fix Procedure, Graphical User Interface |
SVAPAR-137244 |
All |
Suggested
|
In rare circumstances, an internal issue with the GUI backend sorting algorithm can display the following error - 'An error occurred while loading table data'
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
Any volume name containing numbers between dashes, or similar volume names that contain numbers after dashes, but with other characters, such as underscores after these numbers (example - Volume8_Data) |
Workaround |
None |
|
8.6.0.5 |
GUI Fix Procedure, Graphical User Interface |
SVAPAR-137906 |
All |
Suggested
|
A node warmstart may occur due to a timeout caused by FlashCopy bitmap cleaning, leading to a stalled software upgrade.
(show details)
Symptom |
Single Node Warmstart |
Environment |
This can occur on any platform configured with FlashCopy, but it is much more likely to cause an outage on a low end, low core systems. |
Trigger |
System upgrade |
Workaround |
Set the clean rate of all the maps to 0 before upgrading, and then reverting them back again afterwards. |
|
8.6.0.5 |
FlashCopy, System Update |
SVAPAR-138418 |
All |
Suggested
|
Snap collections triggered by Storage Insights over cloud callhome time out before they have completed
(show details)
Symptom |
None |
Environment |
Systems using Cloud Callhome |
Trigger |
Remotely collecting a snap via callhome |
Workaround |
Collect the snap locally and upload to IBM |
|
8.6.0.5 |
Reliability Availability Serviceability |
SVAPAR-140994 |
All |
Suggested
|
Expanding a volume via the GUI fails with CMMVC7019E because the volume size is not a multiple of 512 bytes.
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
Attempting to expand a volume by a value that is not a multiple of 512 bytes |
Workaround |
Expand the volume via the CLI using the 'expandvdisksize' command, ensuring that you increase by 512 byte multiples |
|
8.6.0.5 |
Reliability Availability Serviceability |
SVAPAR-141001 |
All |
Suggested
|
Unexpected error CMMVC9326E when adding either a port to host or creating a host.
(show details)
Symptom |
Configuration |
Environment |
Systems containing less than four I/O groups that contain more than 512 hosts per I/O group |
Trigger |
Adding a port to a host, or creating a host. |
Workaround |
None |
|
8.6.0.5 |
Hosts |
SVAPAR-141019 |
All |
Suggested
|
The GUI crashed when a user group with roles 3SiteAdmin and remote users exist
(show details)
Symptom |
Configuration |
Environment |
Systems configured for 3Ssite replication |
Trigger |
Environments that have a user group with role 3SiteAdmin and a remote user exists |
Workaround |
Either remove the remote user, or if no 3 site replication is in use, remove the user group with role 3SiteAdmin |
|
8.6.0.5 |
3-Site using HyperSwap or Metro Mirror, Graphical User Interface |
SVAPAR-116592 |
All |
HIPER
|
If a V5000E or a Flashsystem 5000 is configured with multiple compressed IP partnerships, and one or more of the partnerships is with a non V5000E or FS5000, it may repeatedly warmstart due to a lack of compression resources.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
V5000E or Flashsystem 5000 configured with multiple compressed IP partnerships, and one or more of the partnerships is with a non V5000E or FS5000. Note a Flashsystem 5200 is not included in FS5000 here. |
Trigger |
Configuring a V5000E or Flashsystem 5000 with multiple compressed IP partnerships, and one or more of the partnerships is with a non V5000E or FS5000. |
Workaround |
Turn off compression for the partnership with the non V5000E or FS5000 system. |
|
8.6.0.4 |
IP Replication |
SVAPAR-131567 |
FS7300, FS9500, SVC |
HIPER
|
Node goes offline and enters service state when collecting diagnostic data for 100Gb/s adapters.
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems that have 100Gb/s adapters installed. |
Trigger |
Collecting support package with statesaves. |
Workaround |
Avoid collecting a support package with statesaves |
|
8.6.0.4 |
|
SVAPAR-132123 |
All |
HIPER
|
Vdisks can go offline after a T3 with an expanding DRAID1 array evokes some IO errors and data corruption
(show details)
Symptom |
Offline Volumes |
Environment |
Any cluster running on 8.5.0.0 or 8.6.x.x code and above. |
Trigger |
When we inject a T3 disaster recovery with an expanding DRAID1 array we get the offline vdisk and the data corrupted and/or IO errors. |
Workaround |
None |
|
8.6.0.4 |
RAID |
SVAPAR-111444 |
All |
Critical
|
Direct attached fibre channel hosts may not log into the NPIV host port due to a timing issue with the Registered State Change Notification (RSCN).
(show details)
Symptom |
Loss of Access to Data |
Environment |
Direct attached fibre channel hosts where the storage has NPIV enabled. |
Trigger |
Storage node reboot |
Workaround |
None |
|
8.6.0.4 |
Host Cluster, Hosts |
SVAPAR-112939 |
All |
Critical
|
A loss of disk access on one pool may cause IO to hang on a different pool due to a cache messaging hang.
(show details)
Symptom |
Loss of Access to Data |
Environment |
System with multiple storage pools. |
Trigger |
Loss of disk access to one pool. |
Workaround |
None |
|
8.6.0.4 |
Cache |
SVAPAR-115478 |
FS7300 |
Critical
|
An issue in the thin-provisioning component may cause a node warmstart during upgrade from pre-8.5.4 to 8.5.4 or later.
(show details)
Symptom |
Single Node Warmstart |
Environment |
FS7300 systems running levels lower then 8.5.4.0 |
Trigger |
Upgrading from pre 8.5.4.0 to to 8.5.4.0 or later |
Workaround |
None |
|
8.6.0.4 |
Thin Provisioning |
SVAPAR-115505 |
All |
Critical
|
Expanding a volume in a Flashcopy map and then creating a dependent incremental forward and reverse Flashcopy map may cause a dual node warmstart when the incremental map is started.
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using incremental reverse Flashcopy mappings. |
Trigger |
Expanding a volume in a Flashcopy map and then creating and starting a dependent incremental forward and reverse Flashcopy map. |
Workaround |
None |
|
8.6.0.4 |
FlashCopy |
SVAPAR-123874 |
All |
Critical
|
There is a timing window when using async-PBR or RC GMCV, with Volume Group snapshots, which results in the new snapshot VDisk mistakenly being taken offline, forcing the production volume offline for a brief period.
(show details)
Symptom |
Offline Volumes |
Environment |
Policy-based Replication or Global Mirror with Change Volumes |
Trigger |
When a snapshot begins, and at the same moment the change volume mappings are being prepared and about to trigger. |
Workaround |
None |
|
8.6.0.4 |
Global Mirror With Change Volumes, Policy-based Replication |
SVAPAR-123945 |
All |
Critical
|
If a system SSL certificate is installed with the extension CA True it may trigger multiple node warmstarts.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
None |
Trigger |
Installing a system SSL certificate with the extension CA True |
Workaround |
None |
|
8.6.0.4 |
Encryption |
SVAPAR-126767 |
All |
Critical
|
Upgrading to 8.6.0 when iSER clustering is configured, may cause multiple node warmstarts to occur, if node canisters have been swapped between slots since the system was manufactured.
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using iSER clustering, where node canisters have been swapped. |
Trigger |
Upgrading to 8.6.0.0 when iSER clustering is configured |
Workaround |
None |
|
8.6.0.4 |
iSCSI |
SVAPAR-127836 |
All |
Critical
|
Running some Safeguarded Copy commands can cause a cluster recovery in some platforms.
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using either Safeguarded copy 1.0 or 2.0 |
Trigger |
Running SGC 1.0 commands on platforms that only support SGC 2.0 or vice versa can cause a cluster recovery. |
Workaround |
Do not run non-supported SGC related commands |
|
8.6.0.4 |
Safeguarded Copy & Safeguarded Snapshots |
SVAPAR-128052 |
All |
Critical
|
A node assert may occur if a host sends a login request to a node when the host is being removed from the cluster with the '-force' parameter.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any host that uses the NVME protocol |
Trigger |
Removing a node by using the '-force' parameter |
Workaround |
Do not use the '-force' parameter |
|
8.6.0.4 |
Hosts, NVMe |
SVAPAR-128626 |
All |
Critical
|
A node may warmstart or fail to start FlashCopy maps, in volume groups that contain Remote Copy primary and secondary volumes, or both copies of a Hyperswap volume.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any systems configured with volume groups containing both Remote Copy primary and secondary volumes, or both copies of a Hyperswap volume. |
Trigger |
Starting a FlashCopy map in an affected volume group |
Workaround |
None |
|
8.6.0.4 |
FlashCopy, Global Mirror, Global Mirror With Change Volumes, HyperSwap, Metro Mirror |
SVAPAR-129298 |
All |
Critical
|
Manage disk group went offline during queueing of fibre rings on the overflow list causing the node to assert.
(show details)
Symptom |
Single Node Warmstart |
Environment |
None |
Trigger |
None |
Workaround |
None |
|
8.6.0.4 |
RAID |
SVAPAR-93709 |
FS9500 |
Critical
|
A problem with NVMe drives may impact node to node communication over the PCIe bus. This may lead to a temporary array offline.
(show details)
Symptom |
Loss of Access to Data |
Environment |
None |
Trigger |
Failing NVMe drive |
Workaround |
None |
|
8.6.0.4 |
Drives, RAID |
SVAPAR-108715 |
All |
High Importance
|
The Service Assistant GUI on 8.5.0.0 and above incorrectly performs actions on the local node instead of the node selected in the GUI.
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
None |
Workaround |
The work around is to perform Service Assistant GUI actions using the Service Assistant CLI instead. Alternatively, In the Service Assistant GUI, select a node that you are not on to perform a Service Assistant action, then when you submit the command, the action will be performed on the local node instead. |
|
8.6.0.4 |
Graphical User Interface |
SVAPAR-110743 |
All |
High Importance
|
Email becoming stuck in the mail queue caused a delay in the 'upgrade commit was finished' message being sent, therefore causing 3 out of 4 nodes to warmstart, and then rejoin the cluster automatically within less than three minutes.
(show details)
Symptom |
Single Node Warmstart |
Environment |
None |
Trigger |
Email stuck in the mail queue causing a delay in the 'upgrade commit was finished' message being sent. |
Workaround |
None |
|
8.6.0.4 |
System Update |
SVAPAR-110765 |
All |
High Importance
|
In a 3-Site configuration, the Config node can be lost if the 'stopfcmap or 'stopfcconsistgrp ' commands are run with the '-force' parameter
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any cluster with 3-Site configuration |
Trigger |
Running 'stopfcmap' or 'stopfcconsistgrp' with '-force' on the flash copy maps with vdisk in 3-site configuration. |
Workaround |
Do not use the '-force' parameter when running either the 'stopfcmap' or 'stopfcconsistgrp' commands. |
|
8.6.0.4 |
3-Site using HyperSwap or Metro Mirror |
SVAPAR-112856 |
All |
High Importance
|
Conversion of Hyperswap volumes to 3 site consistency groups will increase write response time of the Hyperswap volumes.
(show details)
Symptom |
Performance |
Environment |
Any system running Hyperswap and 3-Site |
Trigger |
Conversion of Hyperswap to 3 site consistency groups |
Workaround |
Manually increase rsize of Hyperswap change volumes before conversion to 3 site consistency groups |
|
8.6.0.4 |
3-Site using HyperSwap or Metro Mirror, HyperSwap |
SVAPAR-115021 |
All |
High Importance
|
Software validation checks can trigger a T2 recovery when attempting to move a Hyperswap vdisk into and out of the nocachingiogrp state.
(show details)
Symptom |
Loss of Access to Data |
Environment |
Any system that is configured for hyperswap |
Trigger |
Invoking 'movevdisk' command with the '-nocachingiogrp' flag in a Hyperswap environment |
Workaround |
None |
|
8.6.0.4 |
HyperSwap |
SVAPAR-127063 |
All |
High Importance
|
Degraded Remote Copy performance on systems with multiple IO groups running 8.5.0.11 or 8.6.0.3 after a node restarts
(show details)
Symptom |
Performance |
Environment |
Systems with multiple IO groups using Remote Copy |
Trigger |
Restarting a node |
Workaround |
Warmstart any node that is affected by the issue |
|
8.6.0.4 |
Global Mirror, Global Mirror With Change Volumes, HyperSwap, Metro Mirror, Performance |
SVAPAR-127841 |
All |
High Importance
|
A slow I/O resource leak may occur when using FlashCopy, and the system is under high workload. This may cause a node warmstart to occur
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system configured with FlashCopy |
Trigger |
Many FlashCopy activities occurring when the system is experiencing high workloads |
Workaround |
None |
|
8.6.0.4 |
FlashCopy |
SVAPAR-128228 |
All |
High Importance
|
The NTP daemon may not synchronise after upgrading from 8.3.x to 8.5.x
(show details)
Symptom |
None |
Environment |
Any system that has upgraded from 8.3.x to 8.5.x |
Trigger |
Upgrading from 8.3.x to 8.5.x |
Workaround |
None |
|
8.6.0.4 |
|
SVAPAR-130731 |
All |
High Importance
|
During installation, a single node assert at the end of the software upgrade process may occur
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system undergoing code upgrade |
Trigger |
None |
Workaround |
None |
|
8.6.0.4 |
System Update |
SVAPAR-108469 |
All |
Suggested
|
A single node warmstart may occur on nodes configure to use a secured IP partnership
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with a secured IP partnership |
Trigger |
Service Association goes down |
Workaround |
None |
|
8.6.0.4 |
IP Replication |
SVAPAR-111021 |
All |
Suggested
|
Unable to load resource page in GUI if the IO group ID:0 does not have any nodes.
(show details)
Symptom |
None |
Environment |
Any systems that have no nodes in IO group ID:0. |
Trigger |
None |
Workaround |
None |
|
8.6.0.4 |
System Monitoring |
SVAPAR-111992 |
All |
Suggested
|
Unable to configure policy-based Replication using the GUI, if truststore contains blank lines or CRLF line endings
(show details)
Symptom |
Configuration |
Environment |
Systems configured with policy-based Replication |
Trigger |
Attempting to configure policy-based Replication |
Workaround |
Ensure the certificate file used to create the truststore contains no blank lines and uses LF line endings instead of CRLF line endings. |
|
8.6.0.4 |
Graphical User Interface, Policy-based Replication |
SVAPAR-113792 |
All |
Suggested
|
Node assert may occur when outbound IPC message such as nslookup to a DNS server timeouts
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system running 8.6.0.x or higher |
Trigger |
None |
Workaround |
None |
|
8.6.0.4 |
|
SVAPAR-114081 |
All |
Suggested
|
The lsfabric command may show FC port logins which no longer exist. In large environments with many devices attached to the SAN, this may result in an incorrect 1800 error being reported, indicating that a node has too many logins.
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
None |
Workaround |
Warmstart each node in turn, to remove the invalid entries from lsfabric. |
|
8.6.0.4 |
Reliability Availability Serviceability |
SVAPAR-120156 |
FS5000, FS5100, FS5200, FS7200, FS7300, SVC |
Suggested
|
An internal process introduced in 8.6.0 to collect iSCSI port statistics can cause host performance to be affected
(show details)
Symptom |
Single Node Warmstart |
Environment |
8.6.0.x and higher code level with 25Gb ethernet adapters |
Trigger |
Any system installed with 25Gb ethernet adapters that performs VMware clone or vMotion or Windows ODX file copy operations |
Workaround |
None |
|
8.6.0.4 |
Performance, iSCSI |
SVAPAR-120399 |
All |
Suggested
|
A host WWPN incorrectly shows as being still logged into the storage when it is not.
(show details)
Symptom |
Configuration |
Environment |
Systems using Fibre Channel host connections. |
Trigger |
Disabling or removing a host fibre channel connection. |
Workaround |
None |
|
8.6.0.4 |
Reliability Availability Serviceability |
SVAPAR-120495 |
All |
Suggested
|
A node can experience performance degradation, if using the embedded VASA provider, thereby leading to a potential single node warmstart.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running with Embedded VASA provider. |
Trigger |
None |
Workaround |
None |
|
8.6.0.4 |
|
SVAPAR-120610 |
All |
Suggested
|
Excessive 'chfcmap' commands can result in multiple node warmstarts occurring
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any systems configured with flashcopy. |
Trigger |
Performing excessive 'chfcmap' commands |
Workaround |
None |
|
8.6.0.4 |
FlashCopy |
SVAPAR-120639 |
All |
Suggested
|
The vulnerability scanner claims cookies were set without HttpOnly flag.
(show details)
Symptom |
Configuration |
Environment |
On port 442, the secure flag from SSL cookie is not set from SSL cookie and the HttpOnly Flag is not set from the cookie. |
Trigger |
None |
Workaround |
None |
|
8.6.0.4 |
|
SVAPAR-121334 |
All |
Suggested
|
Packets with unexpected size are received on the ethernet interface. This causes the internal buffers to become full, thereby causing a node to warmstart to clear the condition
(show details)
Symptom |
Single Node Warmstart |
Environment |
Effects systems running 8.6.x |
Trigger |
Storage code expects packet of size 128 but the initiator is sending a packet of size 110 which causes the node to warmstart |
Workaround |
None |
|
8.6.0.4 |
NVMe |
SVAPAR-122411 |
All |
Suggested
|
A node may assert when a vdisk has been expanded and rehome has not made aware of the possible change of the number of regions it may have to rehome.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system configure with Data Reduction Pools |
Trigger |
Any command that will expand the size of a vdisk such as 'expandvdisksize'. |
Workaround |
None |
|
8.6.0.4 |
Data Reduction Pools |
SVAPAR-123644 |
All |
Suggested
|
A system with NVMe drives may falsely log an error indicating a Flash drive has high write endurance usage. The error cannot be cleared.
(show details)
Symptom |
Configuration |
Environment |
Systems with NVMe drives |
Trigger |
None |
Workaround |
None |
|
8.6.0.4 |
Reliability Availability Serviceability |
SVAPAR-126742 |
All |
Suggested
|
A 3400 error (too many compression errors) may be logged incorrectly, due to an incorrect threshold. The error can be ignored on code levels which do not contain this fix.
(show details)
Symptom |
Configuration |
Environment |
Systems using DRP compression |
Trigger |
None |
Workaround |
None |
|
8.6.0.4 |
Compression, Data Reduction Pools |
SVAPAR-127844 |
All |
Suggested
|
The user is informed that a snapshot policy cannot be assigned. The error message CMMVC9893E is displayed.
(show details)
Symptom |
None |
Environment |
Any system that is configured to use snapshots |
Trigger |
Assigning a policy to a volume group |
Workaround |
None |
|
8.6.0.4 |
FlashCopy |
SVAPAR-127908 |
All |
Suggested
|
A volume mapped to a NVMe host cannot be mapped to another NVMe host via the GUI, however it is possible via the CLI. In addition, when a host is removed from a host cluster, it is not possible to add it back using the GUI
(show details)
Symptom |
None |
Environment |
Any volume mapped to an NVMe hosts |
Trigger |
Any volume that has been mapped to an NVMe hosts, or any host that is deleted from a host cluster, and then added back to same host cluster using the GUI. |
Workaround |
None |
|
8.6.0.4 |
GUI Fix Procedure, Graphical User Interface, Host Cluster, Hosts, NVMe |
SVAPAR-129111 |
All |
Suggested
|
When using the GUI, the IPV6 field is not wide enough, thereby causing the user to scroll right to see the full IPV6 address.
(show details)
Symptom |
None |
Environment |
None |
Trigger |
None |
Workaround |
Use the CLI instead |
|
8.6.0.4 |
Graphical User Interface |
SVAPAR-130729 |
All |
Suggested
|
When upgrading to 850, remote users configured with public keys do not failback to password prompt, if a key is not available.
(show details)
Symptom |
None |
Environment |
Any system with remote users configured with public keys |
Trigger |
None |
Workaround |
None |
|
8.6.0.4 |
Security |
SVAPAR-107547 |
All |
Critical
|
If there are more than 64 logins to a single Fibre Channel port, and a switch zoning change is made, a single node warmstart may occur.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with Fibre Channel adapters |
Trigger |
Switch zoning change with more than 64 logins to a single storage system port. |
Workaround |
Reduce the number of logins to a single storage system port |
|
8.6.0.3 |
Reliability Availability Serviceability |
SVAPAR-111705 |
All |
Critical
|
If a Volume Group Snapshot fails and the system has 'snapshotpreserveparent' set to 'yes', this may trigger multiple node warmstarts.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using Volume Group Snapshot and have 'snapshotpreserveparent' set to 'yes'. |
Trigger |
An addsnapshot failure. |
Workaround |
None |
|
8.6.0.3 |
FlashCopy |
SVAPAR-112107 |
FS9500, SVC |
Critical
|
There is an issue that affects PSU firmware upgrades in FS9500 and SV3 systems that can cause an outage. This happens when one PSU fails to download the firmware and another PSU starts to download the firmware. It is a very rare timing window that can be triggered if two PSUs are reseated close in time during the firmware upgrade process.
(show details)
Symptom |
Loss of Access to Data |
Environment |
FS9500 or SV3 |
Trigger |
Two PSUs are reseated close in time during the firmware upgrade process |
Workaround |
None |
|
8.6.0.3 |
System Update |
SVAPAR-112707 |
SVC |
Critical
|
Marking error 3015 as fixed on a SVC cluster containing SV3 nodes may cause a loss of access to data. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems containing 214x-SV3 nodes that have been downgraded from 8.6 to 8.5 |
Trigger |
Marking 3015 error as fixed |
Workaround |
Do not attempt to repair the 3015 error, contact IBM support |
|
8.6.0.3 |
Reliability Availability Serviceability |
SVAPAR-115136 |
FS5100, FS5200, FS7200, FS7300, FS9100, FS9200, FS9500 |
Critical
|
Failure of an NVMe drive has a small probability of triggering a PCIe credit timeout in a node canister, causing the node to reboot.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with NVMe drives |
Trigger |
Drive failure |
Workaround |
None |
|
8.6.0.3 |
Drives |
IT41447 |
All |
High Importance
|
When removing the DNS server configuration, a node may discover unexpected metadata and warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running 8.5 or higher |
Trigger |
Removal of DNS server configuration |
Workaround |
None |
|
8.6.0.3 |
Reliability Availability Serviceability |
SVAPAR-110426 |
All |
High Importance
|
When a security admin other than superuser runs security patch related commands 'lspatch' and 'lssystempatches' this can cause a node to warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
None |
Trigger |
Any user that is not Superuser running a security patch command |
Workaround |
Use superuser for security patching commands. |
|
8.6.0.3 |
Security |
SVAPAR-110819 & SVAPAR-113122 |
All |
High Importance
|
A single-node warmstart may occur when a Fibre Channel port is disconnected from one fabric, and added to another. This is caused by a timing window in the FDMI discovery process.
(show details)
Symptom |
Single Node Warmstart |
Environment |
There are more than one fc switches connected to same Flash system cluster |
Trigger |
Disconnected a Fibre Channel port from one switch, and connecting it to another switch. |
Workaround |
None |
|
8.6.0.3 |
Fibre Channel |
SVAPAR-111812 |
All |
High Importance
|
Systems with 8.6.0 or later software may fail to complete lsvdisk commands, if a single SSH session runs multiple lsvdisk commands piped to each other. This can lead to failed login attempts for the GUI and CLI, and is is more likely to occur if the system has more than 400 volumes.
(show details)
Symptom |
Configuration |
Environment |
Systems with 8.6.0 or later software. |
Trigger |
Unusual use of nested svcinfo commands on the CLI. |
Workaround |
Avoid nested svcinfo commands. |
|
8.6.0.3 |
Command Line Interface |
SVAPAR-112525 |
All |
High Importance
|
A node assert can occur due to a resource allocation issue in a small timing window when using Remote Copy
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Hyperswap, Metro Mirror, Global Mirror, or Global Mirror with Change Volumes |
Trigger |
None |
Workaround |
None |
|
8.6.0.3 |
Global Mirror, Global Mirror With Change Volumes, HyperSwap, Metro Mirror |
SVAPAR-117768 |
All |
High Importance
|
Cloud Callhome may stop working without logging an error
(show details)
Symptom |
Configuration |
Environment |
8.6.0 or higher Systems sending data to Storage Insights without using the data collector are most likely to hit this issue |
Trigger |
None |
Workaround |
Cloud callhome can be disabled then re-enabled to restart the callhome if it has failed. |
|
8.6.0.3 |
Call Home |
SVAPAR-102382 |
All |
Suggested
|
Fibre Channel Read Diagnostic Parameters (RDP) indicates that a short wave SFP is installed when infact an long wave SFP is installed.
(show details)
Symptom |
Configuration |
Environment |
Systems with long wave SFPs running 8.5.2.0 or higher. |
Trigger |
None |
Workaround |
None |
|
8.6.0.3 |
System Monitoring |
SVAPAR-105955 |
All |
Suggested
|
Single node warmstart during link recovery when using a secured IP partnership.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using a secured IP partnership. |
Trigger |
IP link disconnection and reconnection. |
Workaround |
None |
|
8.6.0.3 |
IP Replication |
SVAPAR-108551 |
All |
Suggested
|
An expired token in the GUI file upload process can cause the upgrade to not start automatically after the file is successfully uploaded.
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
Start a code upgrade via GUI |
Workaround |
Upgrade via the CLI. The other option is to log out from the GUI, then log back in to re-authenticate, then go back to the Update System view. On the Test and Update modal, select the test utility and update image files that are already on the system from the previous upload (without selecting to upload them again) |
|
8.6.0.3 |
System Update |
SVAPAR-112711 |
All |
Suggested
|
IBM Storage Virtualize user interface code will not respond to a malformed HTTP POST with expected HTTP 401 message.
(show details)
Symptom |
None |
Environment |
IBM Storage Virtualize GUI |
Trigger |
Malformed HTTP POST |
Workaround |
None |
|
8.6.0.3 |
Graphical User Interface |
SVAPAR-112712 |
SVC |
Suggested
|
The Cloud Call Home function will not restart on SVC clusters that were initially created with CG8 hardware and upgraded to 8.6.0.0 and above.
(show details)
Symptom |
None |
Environment |
SVC cluster that has been upgraded from CG8 hardware. |
Trigger |
Upgrading SVC cluster |
Workaround |
None |
|
8.6.0.3 |
Call Home |
SVAPAR-117179 |
All |
Suggested
|
Snap data collection does not collect an error log if the superuser password requires a change
(show details)
Symptom |
None |
Environment |
None |
Trigger |
Collecting a snap after the superuser password has expired. |
Workaround |
Change the superuser password |
|
8.6.0.3 |
Support Data Collection |
SVAPAR-117781 |
All |
Suggested
|
A single node warmstart may occur during Fabric Device Management Interface (FDMI) discovery if a virtual WWPN is discovered on a different physical port than is was previously.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running 8.5.2 or higher and other devices on the same Fibre Channel fabric using virtual WWPNs. |
Trigger |
Moving a virtual WWPN to a different physical port. |
Workaround |
None |
|
8.6.0.3 |
Hosts |
SVAPAR-105861 |
SVC |
HIPER
|
A cluster recovery may occur when an attempt is made to create a mirrored snapshot with insufficient volume mirroring bitmap space in the IO group
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using mirrored snapshots, including safeguarded snapshots |
Trigger |
Attempting to create a mirrored snapshot without sufficient volume mirroring bitmap space |
Workaround |
Adjust snapshot policy configuration to ensure that the maximum bitmap space is sufficient |
|
8.6.0.2 |
FlashCopy, Safeguarded Copy & Safeguarded Snapshots, Volume Mirroring |
SVAPAR-104533 |
All |
Critical
|
Systems that encounter multiple node asserts, followed by a system T3 recovery, may experience errors repairing Data Reduction Pools
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools |
Trigger |
Multiple node asserts followed by a system T3 recovery |
Workaround |
None |
|
8.6.0.2 |
Data Reduction Pools |
SVAPAR-105430 |
All |
Critical
|
When hardware compression is suspended mid IO to a DRP compressed volume, it may cause the IO to hang until an internal timeout is hit and a node warmstarts.
(show details)
Symptom |
Single Node Warmstart |
Environment |
GEN3 or later hardware with DRP compressed volumes. |
Trigger |
None |
Workaround |
None |
|
8.6.0.2 |
Compression, Data Reduction Pools |
SVAPAR-107270 |
All |
Critical
|
If an upgrade from a level below 8.6.x, to 8.6.0 or 8.6.1 commits, whilst FlashCopy is preparing to start a map, a bad state is introduced that prevents the FlashCopy maps from starting.
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems using Global Mirror with Change Volumes or Policy-based Replication |
Trigger |
Upgrade commit |
Workaround |
Stop replication, or the partnership before starting the upgrade |
|
8.6.0.2 |
Global Mirror With Change Volumes, Policy-based Replication |
SVAPAR-107734 |
All |
Critical
|
When issuing IO to an incremental fcmap volume that is in a stopped state, but has recently been expanded, and it also has a partner fcmap, may cause the nodes to go into a restart.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems configured with Incremental flashcopy and reverse fcmap |
Trigger |
Resizing volumes in an incremental partnered fcmaps |
Workaround |
Ensure that both incremental partnered fcmaps are deleted, and then re-create a new pair if you need to resize the volumes. |
|
8.6.0.2 |
FlashCopy |
SVAPAR-114899 |
All |
Critical
|
Out of order snapshot stopping can cause stuck cleaning processes to occur, following Policy-based Replication cycling. This manifests as extremely high CPU utilization on multiple CPU cores, causing excessively high volume response times.
(show details)
Symptom |
Performance |
Environment |
Any system configured with Policy-based Replication and Volume group snapshots |
Trigger |
Snapshots are created. A volume group then switches to Policy-based Replication cycling mode but remains in journalling mode. Dependent snapshots are removed which triggers cleaning. The affected snapshots remain in stopping state even after cleaning completes. |
Workaround |
Check for dependent vdisks and reset the nodes one at a time using 'svctask stopsystem -reset -node x'. Allow sufficient time for multipathing to recover between each node restart. |
|
8.6.0.2 |
Policy-based Replication |
SVAPAR-104159 |
All |
High Importance
|
Nodes configured with 32GB or less of RAM, and specific 25Gb ethernet adapters, under some circumstances may run out of memory. This can cause a single node warmstart.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Nodes running 8.6.0.0 with 32GB or less of RAM, and specific 25Gb ethernet adapters |
Trigger |
None |
Workaround |
None |
|
8.6.0.2 |
Reliability Availability Serviceability |
SVAPAR-104250 |
All |
High Importance
|
There is an issue whereby NVMe CaW (Compare and Write) commands can incorrectly go into an invalid state, thereby causing the node to assert to clear the bad condition
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any host running NVMe |
Trigger |
None |
Workaround |
None |
|
8.6.0.2 |
Hosts, NVMe |
SVAPAR-105727 |
All |
High Importance
|
An upgrade within the 8.5.0 release stream from 8.5.0.5 or below, to 8.5.0.6 or above, can cause an assert of down-level nodes during the upgrade, if volume mirroring is heavily utilised
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any system running volume mirroring with either a large number of volumes or high syncrate |
Trigger |
Upgrading from 8.5.0.5 or below to 8.5.0.6 or above with heavy volume mirroring workload |
Workaround |
Disable mirroring, or reduce syncrate to a low value during the upgrade process |
|
8.6.0.2 |
Volume Mirroring |
SVAPAR-106874 |
FS5200, FS7200, FS7300, FS9100, FS9200, FS9500, SVC |
High Importance
|
A timing window may cause a single node warmstart, while recording debug information about a replicated host write. This can only happen on a system using Policy Based Replication.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running PBR and code 8.6.0.0 or 8.6.0.1 |
Trigger |
None |
Workaround |
None |
|
8.6.0.2 |
Policy-based Replication |
SVAPAR-99997 |
All |
High Importance
|
Creating a volume group from a snapshot whose index is greater than 255 may cause incorrect output from 'lsvolumegroup'
(show details)
Symptom |
Configuration |
Environment |
Systems using Volume Group Snapshots |
Trigger |
Creating a volume group from a snapshot whose index is greater than 255 |
Workaround |
None |
|
8.6.0.2 |
FlashCopy |
SVAPAR-102271 |
All |
Suggested
|
Enable IBM Storage Defender integration for Data Reduction Pools
(show details)
Symptom |
None |
Environment |
None |
Trigger |
None |
Workaround |
None |
|
8.6.0.2 |
Interoperability |
SVAPAR-106693 |
FS9500 |
Suggested
|
Remote Support Assistance (RSA) cannot be enabled on FS9500 systems with MTM 4983-AH8
(show details)
Symptom |
Configuration |
Environment |
FS9500 systems with MTM 4983-AH8 |
Trigger |
Trying to enable RSA on MTM 4983-AH8 |
Workaround |
None |
|
8.6.0.2 |
Support Remote Assist |
SVAPAR-107558 |
All |
Suggested
|
A Volume Group Snapshot (VGS) trigger may collide with a GMCV or Policy based Replication cycle causing the VGS trigger to fail.
(show details)
Symptom |
Configuration |
Environment |
Volume Group Snapshots with GMCV or Policy Based Replication |
Trigger |
Trigger a Volume Group Snapshot |
Workaround |
None |
|
8.6.0.2 |
FlashCopy, Global Mirror With Change Volumes, Policy-based Replication |
SVAPAR-107595 |
FS7300, FS9100, FS9200, FS9500, SVC |
Suggested
|
Improve maximum throughput for Global Mirror, Metro Mirror and Hyperswap by providing more inter-node messaging resources
(show details)
Symptom |
Performance |
Environment |
Systems running Global Mirror, Metro Mirror or Hyperswap |
Trigger |
High Global Mirror, Metro Mirror or Hyperswap workload |
Workaround |
None |
|
8.6.0.2 |
Global Mirror, HyperSwap, Metro Mirror, Performance |
SVAPAR-107733 |
All |
Suggested
|
The 'mksnmpserver' command fails with 'CMMVC5711E [####] is not valid data' if auth passphrase contains special characters, such as '!'
(show details)
Symptom |
Configuration |
Environment |
Any system running on or after v8.3.1.9, v8.4.0.10, or v8.5.0.7 |
Trigger |
Using an auth passphrase containing special characters to execute the 'mksnmpserver' command |
Workaround |
Do not include special characters in the auth passphrase |
|
8.6.0.2 |
|
SVAPAR-109289 |
All |
Suggested
|
Buffer overflow may occur when handling the maximum length of 55 characters for either Multi-Factor Authentication (MFA) or Single Sign On (SSO) client secrets
(show details)
Symptom |
Configuration |
Environment |
Systems that use MFA or SSO |
Trigger |
Using a client secret with > 55 characters |
Workaround |
Use less then 55 characters for the client secret |
|
8.6.0.2 |
Backend Storage |
SVAPAR-98576 |
All |
Suggested
|
Customers cannot edit certain properties of a flashcopy mapping via the GUI flashcopy mappings panel as the edit modal does not appear.
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
None |
Workaround |
Use the CLI instead |
|
8.6.0.2 |
FlashCopy, Graphical User Interface |
SVAPAR-103696 |
All |
HIPER
|
When taking a snapshot of a volume that is being replicated to another system using Policy Based Replication, the snapshot may contain data from an earlier point in time than intended
(show details)
Symptom |
None |
Environment |
Systems using snapshots of Policy Based Replication production volumes, where the replication is running in cycling mode |
Trigger |
Creating a snapshot |
Workaround |
None |
|
8.6.0.1 |
FlashCopy, Policy-based Replication |
SVAPAR-94179 |
FS5100, FS5200, FS7200, FS7300, FS9100, FS9200, FS9500 |
HIPER
|
Faulty hardware within or connected to the CPU can result in a reboot on the affected node. However it is possible for this to sometimes result in a reboot on the partner node
(show details)
Symptom |
Loss of Access to Data |
Environment |
All Flashsystems and V7000 Gen3, but not SVC |
Trigger |
Node hardware fault |
Workaround |
None |
|
8.6.0.1 |
Reliability Availability Serviceability |
HU02585 |
All |
Critical
|
An unstable connection between the Storage Virtualize system and an external virtualized storage system can sometimes result in a cluster recovery occurring
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
None |
Trigger |
An unstable connection between the Storage Virtualize system and an external virtualized storage system can cause objects to be discovered out of order, resulting in a cluster recovery |
Workaround |
Stabilise the SAN fabric by replacing any failing hardware, such as a faulty SFP |
|
8.6.0.1 |
Backend Storage |
SVAPAR-100127 |
All |
Critical
|
The Service Assistant GUI Node rescue option incorrectly performs the node rescue on the local node instead of the node selected in the GUI.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any cluster running on 8.5.0.0 code and above. |
Trigger |
This problem can happen if the user is on the Service Assistant GUI of a node but selects another node for node rescue. The Node rescue will perform on the local node they are on and not the node selected |
Workaround |
Use the CLI 'satask rescuenode -force <node-panel-id>' command to select the correct node to perform the node rescue on, or log onto the Service GUI of the node that is requiring a node rescue if it is accessible, that way the node in need will be the local node |
|
8.6.0.1 |
Graphical User Interface |
SVAPAR-100564 |
All |
Critical
|
On code level 8.6.0.0, multiple node warmstarts will occur if a user attempts to remove the site ID from a host that has Hyperswap volumes mapped to it.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Hyperswap cluster on 8.6.0.0 with Hyperswap volumes mapped to one or more hosts |
Trigger |
Attempting to remove the site ID from a host that has Hyperswap volumes mapped to it |
Workaround |
Convert all the mapped Hyperswap volumes to basic volumes, then remove the site ID |
|
8.6.0.1 |
HyperSwap |
SVAPAR-98184 |
All |
Critical
|
When a Volume Group Snapshot clone is added to a replication policy before the clone is complete, the system may repeatedly warmstart when the Policy-based Replication volume group is changed to independent access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Volume Group Snapshot Clones and Policy-Based Replication |
Trigger |
Changing an affected Policy-Based Replication Volume Group to independent access |
Workaround |
Wait for the clone to complete before adding the volumes to a replication policy |
|
8.6.0.1 |
FlashCopy, Policy-based Replication |
SVAPAR-98612 |
All |
Critical
|
Creating a volume group snapshot with an invalid I/O group value may trigger multiple node warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using volume group snapshots |
Trigger |
Using an invalid I/O group value when creating a volume group snapshot |
Workaround |
Make sure that you specify the correct I/O group value |
|
8.6.0.1 |
FlashCopy |
SVAPAR-98672 |
All |
Critical
|
VMWare host crashes on servers connected using NVMe over Fibre Channel with the host_unmap setting disabled
(show details)
Symptom |
Loss of Access to Data |
Environment |
ESXi hosts connected to a system using NVME over FC protocol |
Trigger |
VM is sending unmap command with type deallocate |
Workaround |
Enabling host unmap will resolve the VMWare instability. However enabling unmap can cause performance issues for systems with enterprise or nearline drives |
|
8.6.0.1 |
NVMe |
SVAPAR-100162 |
All |
High Importance
|
Some host vendors such as Windows have recently started to use 'mode select page 7'. IBM Storage does not support this mode. If the storage received this mode level it would cause a warmstart to occur
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any cluster running on 8.4.0.0 or higher |
Trigger |
If a host uses 'mode select page 7' |
Workaround |
None |
|
8.6.0.1 |
Hosts |
SVAPAR-100977 |
All |
High Importance
|
When a zone containing NVMe devices is enabled, a node warmstart might occur.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system running 8.5.0.5 |
Trigger |
Enabling a zone with a host that has approximately 1,000 vdisks mapped |
Workaround |
Make sure that the created zone does not contain NVMe devices |
|
8.6.0.1 |
NVMe |
SVAPAR-102573 |
All |
High Importance
|
On systems using Policy-Based Replication and Volume Group Snapshots, some CPU cores may have high utilization due to an issue with the snapshot cleaning algorithm. This can impact performance for replication and host I/O
(show details)
Symptom |
Performance |
Environment |
Systems using Policy-Based Replication and Volume Group Snapshots |
Trigger |
Snapshot mappings with low cleaning workload |
Workaround |
None |
|
8.6.0.1 |
Policy-based Replication |
SVAPAR-98497 |
All |
High Importance
|
Excessive SSH logging may cause the Configuration node boot drive to become full. The node will go offline with error 565, indicating a boot drive failure
(show details)
Symptom |
Configuration |
Environment |
Any system that is being monitored by an external monitoring systems |
Trigger |
Customers using external monitoring systems such as Zabbix that use SSH to log in multiple times a second maybe effected |
Workaround |
None |
|
8.6.0.1 |
System Monitoring |
SVAPAR-98893 |
All |
High Importance
|
If an external storage controller has over-provisioned storage (for example a FlashSystem with an FCM array), the system may incorrectly display usable capacity data for mdisks from that controller. If connectivity to the storage controller is lost, node warmstarts may occur
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running 8.6.0.0 only, with an external controller that has over-provisioned storage |
Trigger |
None |
Workaround |
None |
|
8.6.0.1 |
Storage Virtualisation |
SVAPAR-99175 |
All |
High Importance
|
A node may warmstart due to an invalid queuing mechanism in cache. This can cause IO in cache to be in the same processing queue more than once.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any cluster on code below 8.6.1 |
Trigger |
Can happen when IO in cache is being processed |
Workaround |
None |
|
8.6.0.1 |
Cache |
SVAPAR-99354 |
All |
High Importance
|
Missing policing in the 'startfcconsistgrp' command for volumes using volume group snapshots, resulting in node warmstarts when creating a new volume group snapshot
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any system configured with Volume Group Snapshot and FlashCopy |
Trigger |
Adding a Volume Group Snapshot when a downstream legacy FlashCopy map exists. |
Workaround |
Only create a Volume Group Snapshot if a downstream legacy FlashCopy map does not exist |
|
8.6.0.1 |
FlashCopy |
SVAPAR-99537 |
All |
High Importance
|
If a hyperswap volume copy is created in a DRP child pool, and the parent pool has FCM storage, the change volumes will be created as thin-provisioned instead of compressed
(show details)
Symptom |
Configuration |
Environment |
Systems with DRP child pools and FCM storage |
Trigger |
Creating a change volume in a DRP child pool when the parent pool contains FCMs |
Workaround |
None |
|
8.6.0.1 |
Data Reduction Pools |
SVAPAR-99855 |
FS9500, SVC |
High Importance
|
After battery firmware is upgraded on SV3 or FS9500 as part of a software upgrade, there is a small probability that the battery may remain permanently offline
(show details)
Symptom |
Loss of Redundancy |
Environment |
FS9500 or SV3 |
Trigger |
None |
Workaround |
None |
|
8.6.0.1 |
|
SVAPAR-100172 |
FS9500, SVC |
Suggested
|
During the enclosure component upgrade, which occurs after the cluster upgrade has committed, a system can experience spurious 'The PSU has indicated DC failure' events (error code 1126 ). The event will automatically fix itself after several seconds and there is no user action required
(show details)
Symptom |
None |
Environment |
Any FS9500 or SVC SV3 upgrading to 8.6.0.0 |
Trigger |
Any FS9500 or SVC SV3 system that has been upgraded to 8.6.0.0 |
Workaround |
None |
|
8.6.0.1 |
|
SVAPAR-100958 |
All |
Suggested
|
A single FCM may incorrectly report multiple medium errors for the same LBA
(show details)
Symptom |
Performance |
Environment |
Predominantly FCM2, but could also affect other FCM generations |
Trigger |
None |
Workaround |
After the problem is detected, manually fail the FCM, format it and then insert back into the array. After the copyback has completed ensure to update all FCMs to the recommended firmware level |
|
8.6.0.1 |
RAID |
SVAPAR-110059 |
All |
Suggested
|
When using Storage Insights without a data collector, an attempt to collect a snap using Storage Insights may fail.
(show details)
Symptom |
None |
Environment |
Storage Virtualize system running 8.5.4 or higher and using Storage Insights without a data collector |
Trigger |
Initiating a Support Package collection using Storage Insights |
Workaround |
Disable then re-enble the Cloud Callhome service using the 'svctask chcloudcallhome -disable' and 'svctask chcloudcallhome -enable'. This will function until the next time the configuration node fails over |
|
8.6.0.1 |
Support Data Collection |
SVAPAR-95384 |
All |
Suggested
|
In very rare circumstances, a timing window may cause a single node warmstart when creating a volume using policy-based replication
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system configured with Policy-Based Replication |
Trigger |
Running the 'mkvolume' command |
Workaround |
None |
|
8.6.0.1 |
Policy-based Replication |
SVAPAR-97502 |
All |
Suggested
|
Configurations that use Policy-based Replication with standard pool change volumes will raise space usage warnings
(show details)
Symptom |
None |
Environment |
This issue can only be triggered when using Policy-based Replication with standard pools. The issue is not presented within DRP environments |
Trigger |
System that use Policy-based Replication within a standard pool whilst running 8.5.2.0 - 8.6.0.0 |
Workaround |
None |
|
8.6.0.1 |
Policy-based Replication |
SVAPAR-98128 |
All |
Suggested
|
A single node warmstart may occur on upgrade to 8.6.0.0, on SA2 nodes with 25Gb ethernet adapters
(show details)
Symptom |
Single Node Warmstart |
Environment |
SA2 nodes with a 25Gb ethernet adapters |
Trigger |
Upgrading to 8.6.0.0 |
Workaround |
None |
|
8.6.0.1 |
System Update |
SVAPAR-98611 |
All |
Suggested
|
The system returns an incorrect retry delay timer for a SCSI BUSY status response to AIX hosts when an attempt is made to access a VDisk that is not mapped to the host
(show details)
Symptom |
Loss of Access to Data |
Environment |
AIX hosts |
Trigger |
Trying to access an unmapped VDisk from an AIX host |
Workaround |
None |
|
8.6.0.1 |
Interoperability |
HU02475 |
All |
HIPER
|
Power outage can cause reboots on nodes with 25Gb ethernet adapters, necessitating T3 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Any node that has a 25Gb ethernet adapter installed |
Trigger |
Power outage occurs, causing both nodes to experience a kernel panic, meaning cluster information is lost |
Workaround |
T3 recovery will be required |
|
8.6.0.0 |
Reliability Availability Serviceability |
HU02572 |
All |
HIPER
|
When controllers running specified code levels with SAS storage are power cycled or rebooted, there is a chance that 56 bytes of data will be incorrectly restored into the cache, leading to undetected data corruption. The system will attempt to flush the cache before an upgrade, so this defect is less likely during an upgrade.
(show details)
Symptom |
Data Integrity Loss |
Environment |
Controllers running 8.5.0.0 through 8.5.0.6, 8.5.1, 8.5.2 and 8.5.3 must have SAS storage to be vulnerable to this defect. |
Trigger |
A power cycle or node reboot while the cache is not empty can trigger this defect |
Workaround |
None |
|
8.6.0.0 |
Drives |
SVAPAR-90459 |
All |
HIPER
|
Possible undetected data corruption or multiple node warmstarts if a Traditional FlashCopy Clone of a volume is created before adding Volume Group Snapshots to the volume
(show details)
Symptom |
Data Integrity Loss |
Environment |
Any system running code capable of creating FlashCopy Volume Group Snapshots and Traditional FlashCopy Clones |
Trigger |
If a Traditional FlashCopy Clone is taken of a production volume before Volume Group Snapshots are added to that production volume, it is possible that data in the original Traditional Clone will become corrupt |
Workaround |
To avoid this defect, users can use Volume Group Snapshots. If users need to access the data in these Volume Group Snapshots, the user can create a Volume Group Clone of the Volume Group Snapshot instead of creating a Traditional FlashCopy Clone of the production volume |
|
8.6.0.0 |
FlashCopy |
SVAPAR-98567 |
FS5000 |
HIPER
|
In FS50xx nodes, the TPM may become unresponsive after a number of weeks' runtime. This can lead to encryption or mdisk group CLI commands failing, or in some cases node warmstarts. This issue was partially addressed by SVAPAR-83290, but is fully resolved by this second fix.
(show details)
Symptom |
Loss of Access to Data |
Environment |
FS50xx platforms running either V8.4.0,V8.4.1,V8.4.2,V8.5.0,V8.5.1,V8.5.2,V8.5.3 |
Trigger |
Enabling encryption, or creating an encrypted pool |
Workaround |
Reboot each node in turn. Wait 30 minutes between the two nodes in an I/O group, to allow hosts to failover. Check there are no volumes dependent on the second node before proceeding with the reboot. After all nodes have been rebooted, retry the configuration action, which should now complete successfully |
|
8.6.0.0 |
Encryption |
HU02420 |
All |
Critical
|
During an array copyback it is possible for a memory leak to result in the progress stalling and an node warmstart of all nodes resulting in a temporary loss of access
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running < 8.4 |
Trigger |
Memory leak that causes resources used for the copyback to become depleted |
Workaround |
None |
|
8.6.0.0 |
RAID |
HU02441 & HU02486 |
All |
Critical
|
Safeguarded Copy with DRP can cause node warmstarts and mdisk timeouts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.4.2 or later using Safeguarded Copy with DRP |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Data Reduction Pools, Safeguarded Copy & Safeguarded Snapshots |
HU02471 |
All |
Critical
|
After starting a FlashCopy map with -restore in a graph with a GMCV secondary disk that was stopped with -access there can be a data integrity issue
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems using GMCV |
Trigger |
Given a configuration where there is a GMCV secondary volume A that has a cascade of two FlashCopy maps to volumes B and C (A -> B -> C) The mapping B -> C has been started and GMCV is running Stop the GMCV relationship with -access Start FlashCopy map A -> B with -restore Any I/O to volume A will corrupt data on volume C |
Workaround |
Wait for a GMCV backward map to complete before starting related FlashCopy maps with -restore |
|
8.6.0.0 |
FlashCopy, Global Mirror With Change Volumes |
HU02502 |
All |
Critical
|
On upgrade to v8.4.2 or later with FlashCopy active, a node warmstart can occur, leading to a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.4.2 or later with FlashCopy |
Trigger |
Timing window during upgrade |
Workaround |
None |
|
8.6.0.0 |
FlashCopy |
HU02506 |
All |
Critical
|
On a system where NPIV is disabled or in transitional mode, certain hosts may fail to log in after a node warmstart or reboot (for example during an upgrade), leading to loss of access.
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems where NPIV is not enabled. |
Trigger |
Upgrade from pre-8.5 to 8.5 software or later |
Workaround |
None |
|
8.6.0.0 |
Hosts |
HU02519 & HU02520 |
All |
Critical
|
Safeguarded copy source vdisks go offline when its mappings and target vdisks are deleted then recreated in rapid succession
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.4.2 or later |
Trigger |
target vdisks are deleted then recreated in rapid succession |
Workaround |
There currently is no work around for the issue. The vdisk will be offline for as long as 5 minutes but should come back online on its own |
|
8.6.0.0 |
FlashCopy, Safeguarded Copy & Safeguarded Snapshots |
HU02540 |
All |
Critical
|
Deleting a HyperSwap volume copy with dependent Flashcopy mappings can trigger repeated node warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with FlashCopy on HyperSwap volumes |
Trigger |
Deleting a HyperSwap volume copy with dependent Flashcopy mappings |
Workaround |
None |
|
8.6.0.0 |
FlashCopy, HyperSwap |
HU02541 |
All |
Critical
|
In some circumstances, the deduplication replay process on a data reduction pool can become stuck. During this process, IO to the pool is quiesced and must wait for the replay to complete. Because it does not complete, IO to the entire storage pool hangs, which can eventually lead to a loss of access to data.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running DRP - 8.4.2, 8.5+ |
Trigger |
None |
Workaround |
Warmstart the nodes |
|
8.6.0.0 |
Data Reduction Pools, Deduplication |
HU02546 |
FS5200, FS7200, FS7300, FS9100, FS9200, FS9500, SVC |
Critical
|
On systems running 8.5.2.1, and with Policy-based replication configured, if you created more than 1PB of replicated volumes then this can lead to a loss of hardened data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Any system running 8.5.2.1 and with Policy-based replication configured |
Trigger |
Configuring more than 1PB of capacity using policy-based replication. |
Workaround |
Restrict replicated volume capacity to less than 1PB |
|
8.6.0.0 |
Policy-based Replication |
HU02551 |
All |
Critical
|
When creating multiple volumes with a high mirroring sync rate, an node warmstart maybe triggered due to internal resource constraints
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running 8.5.0, 8.5.1, 8.5.2 |
Trigger |
Creating muliple volumes with high mirroring sync rate |
Workaround |
Lower the sync rate to 100 when creating multiple volumes |
|
8.6.0.0 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02556 |
FS9500, SVC |
Critical
|
In rare circumstances, a FlashSystem 9500 (or SV3) node might be unable to boot, requiring a replacement of the boot drive and TPM
(show details)
Symptom |
None |
Environment |
FS9500 running 8.5.x |
Trigger |
This issue can get triggered during code upgrade |
Workaround |
Replace boot drive and TPM |
|
8.6.0.0 |
Encryption |
HU02561 |
All |
Critical
|
If there are a high number of FC mappings sharing the same target, the internal array that is used to track the FC mapping is mishandled, thereby causing it to overrun. This will cause a cluster wide warmstart to occur
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
System running v8.3.1,v8.4.0,v8.4.1,v8.4.2,v8.5.0 |
Trigger |
Cascaded flashcopy mappings and one of the flashcopy target volumes is the source of 255 flashcopy mappings |
Workaround |
None |
|
8.6.0.0 |
FlashCopy |
HU02563 |
All |
Critical
|
Improve dimm slot identification for memory errors
(show details)
Symptom |
Single Node Warmstart |
Environment |
System running v8.4.0,v8.4.1,v8.4.2,v8.5.0,v8.5.1,v8.5.2 |
Trigger |
Bad memory module |
Workaround |
Work with IBM Support to replace the DIMM that reported the uncorrectable error and replace it |
|
8.6.0.0 |
Reliability Availability Serviceability |
HU02567 |
All |
Critical
|
Due to a low probability timing window, FlashCopy reads can occur indefinitely to an offline Vdisk. This can cause host write delays to flashcopy target volumes that can exceed 6 minutes
(show details)
Symptom |
Loss of Redundancy |
Environment |
All |
Trigger |
Offline vdisk in the flashcopy dependency chain |
Workaround |
None |
|
8.6.0.0 |
FlashCopy |
HU02584 |
All |
Critical
|
If a HyperSwap volume is created with cache disabled in a Data Reduction Pool (DRP), multiple node warmstarts may occur.
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any 8.5.x system running HyperSwap |
Trigger |
Creating a HyperSwap volume with cache disabled in a Data Reduction Pool (DRP) |
Workaround |
None |
|
8.6.0.0 |
Data Reduction Pools, HyperSwap |
HU02586 |
All |
Critical
|
When deleting a safeguarded copy volume which is related to a restore operation and another related volume is offline, the system may warmstart repeatedly
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running safeguarded copy |
Trigger |
Deletion of a safeguarded volume while a restore is in operation |
Workaround |
Return any related offline volumes online |
|
8.6.0.0 |
Safeguarded Copy & Safeguarded Snapshots |
IT41088 |
FS5000, FS5100, FS5200 |
Critical
|
Systems with low memory that have a large number of RAID arrays that are resyncing can cause a system to run out of RAID rebuild control blocks
(show details)
Symptom |
Loss of Access to Data |
Environment |
Low memory systems such as 5015/5035 |
Trigger |
Systems with 64gb or less of cache with resync operations spread across multiple RAID arrays |
Workaround |
None |
|
8.6.0.0 |
RAID |
IT41173 |
FS5200 |
Critical
|
If the temperature sensor in an FS5200 system fails in a particular way, it is possible for drives to be powered off, causing a loss of access to data. This type of temperature sensor failure is very rare.
(show details)
Symptom |
Loss of Access to Data |
Environment |
FS5200 |
Trigger |
Temperature sensor failure |
Workaround |
None |
|
8.6.0.0 |
Reliability Availability Serviceability |
SVAPAR-84116 |
All |
Critical
|
The background delete processing for deduplicated volumes might not operate correctly if the preferred node for a deduplicated volume is changed while a delete is in progress. This can result in data loss which will be detected by the cluster when the data is next accessed
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems running with deduplicated volumes on code levels up to 8.4.0.10, 8.5.0.0 through 8.5.0.7 and 8.5.1 through 8.5.4 are vulnerable to APAR SVAPAR-84116. The fix is available in 8.4.0.11 and 8.5.0.8 |
Trigger |
The background delete processing for deduplicated volumes might not operate correctly if the preferred node for a deduplicated volume is changed while a delete of another volume is in progress. This can result in data loss which will be detected by the cluster when the data is next accessed |
Workaround |
Commands that change the preferred node of a deduplicated volume should not be run while another deduplicated volume is in a deleting state. These commands are: splitvdiskcopy and movevdisk |
|
8.6.0.0 |
Data Reduction Pools, Deduplication |
SVAPAR-86477 |
All |
Critical
|
In some situations ordered processes need to be replayed to ensure the continued management of user workloads. Circumstances exist where this processing can fail to get scheduled so the work remains locked. Software timers that check for this continued activity will detect a stall and force a recovery warmstart
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
None |
Trigger |
Maintenance of nodes (shutting down nodes) triggered a instance of this |
Workaround |
None |
|
8.6.0.0 |
Data Reduction Pools |
SVAPAR-87729 |
All |
Critical
|
After a system has logged '3201 : Unable to send to the cloud callhome servers', the system may end up with an inconsistency in the Event Log. This inconsistency can cause a number of symptoms, including node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any system running V8.4.1,V8.4.2,V8.5.0,V8.5.1,V8.5.2,V8.5.3 |
Trigger |
Cloud callhome errors |
Workaround |
None |
|
8.6.0.0 |
Call Home |
SVAPAR-87846 |
All |
Critical
|
Node warmstarts with unusual workload pattern on volumes with Policy-based replication
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any system running V8.5.2 or V8.5.3 and has Policy-based replication configured |
Trigger |
Unusual workload pattern on volumes with replication policy. Specifically, if a large number (tens of thousands) of writes to the same 128k grain on a single VDisk is submitted in a short amount of time |
Workaround |
Stop the workload pattern, disable replication, or force replication into cycling mode on the affected volume groups |
|
8.6.0.0 |
Policy-based Replication |
SVAPAR-88279 |
All |
Critical
|
A low probability timing window exists in the Fibre Channel login management code. If there are many logins, and two nodes go offline in a very short time, this may cause other nodes in the cluster to warmstart
(show details)
Symptom |
None |
Environment |
Systems with a high number of fibre channel logins (for example, lots of hosts) |
Trigger |
Node warmstart immediately after the partner node has been taken offline |
Workaround |
None |
|
8.6.0.0 |
Reliability Availability Serviceability |
SVAPAR-88887 |
FS9100, FS9200, FS9500 |
Critical
|
Loss of access to data after replacing all boot drives in system
(show details)
Symptom |
Loss of Access to Data |
Environment |
Any FS9xxx cluster |
Trigger |
On a FS9xxx system, if the canisters are swapped, and at some point later, all the boot drives in a canister are replaced |
Workaround |
None |
|
8.6.0.0 |
Drives, Reliability Availability Serviceability |
SVAPAR-89172 |
All |
Critical
|
Snapshot volumes created by running the 'addsnapshot' command from the CLI can be slow to come online, this causes the Production volumes to incorrectly go offline
(show details)
Symptom |
Offline Volumes |
Environment |
Any system running V8.5.2, V8.5.3 or V8.5.4 that uses either FlashCopy or Safeguarded Copy |
Trigger |
Starting Volume Group Snapshots by using the 'addsnapshot' command |
Workaround |
None |
|
8.6.0.0 |
FlashCopy, Safeguarded Copy & Safeguarded Snapshots |
SVAPAR-89692 |
FS9500, SVC |
Critical
|
Battery back-up units may reach end of life prematurely on FS9500 / SV3 systems, despite the batteries being in good physical health, which will result in node errors and potentially nodes going offline if both batteries are affected
(show details)
Symptom |
Loss of Redundancy |
Environment |
FS9500 and SV3 systems are exposed |
Trigger |
There is no trigger. The issue has an increasing likelihood to occur after the batteries are 4-6 months old |
Workaround |
Contact IBM Support to obtain a utility to upgrade the battery firmware |
|
8.6.0.0 |
|
SVAPAR-89764 |
All |
Critical
|
There is an issue with the asynchronous delete behavior of the Safeguarded Copies VDisks in the background that can cause an unexpected internal state in the FlashCopy component that can cause a single node assert
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system that is running Safeguarded Copies |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Safeguarded Copy & Safeguarded Snapshots |
SVAPAR-90438 |
All |
Critical
|
A conflict of host IO on one node, with array resynchronisation task on the partner node, can result in some regions of parity inconsistency. This is due to the asynchronous parity update behaviour leaving invalid parity in the RAID internal cache
(show details)
Symptom |
None |
Environment |
Anything |
Trigger |
This issue can only occur after the RAID array has undergone a re-initialization procedure (such as after a Tier3 recovery) |
Workaround |
None |
|
8.6.0.0 |
Distributed RAID |
SVAPAR-91111 |
All |
Critical
|
USB devices connected to an FS5035 node may be formatted on upgrade to 8.5.3 software
(show details)
Symptom |
Loss of Access to Data |
Environment |
FS5035 with a USB stick connected |
Trigger |
Upgrading to 8.5.3 |
Workaround |
Remove USB device before upgrading to 8.5.3 |
|
8.6.0.0 |
Encryption |
SVAPAR-91860 |
All |
Critical
|
If an upgrade is started with the pause flag and then aborted, the pause flag may not be cleared. This can trigger the system to encounter an unexpected code path on the next upgrade, thereby causing a loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
All SpecV Systems |
Trigger |
Starting an upgrade with the pause flag and then aborting it |
Workaround |
None |
|
8.6.0.0 |
System Update |
SVAPAR-92579 |
All |
Critical
|
If Volume Group Snapshots are in use on a Policy-Based Replication DR system, a timing window may result in a node warmstart for one or both nodes in the I/O group
(show details)
Symptom |
Loss of Access to Data |
Environment |
Policy-Based Replication or GMCV DR clusters with Volume Group Snapshots in use |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Policy-based Replication, Safeguarded Copy & Safeguarded Snapshots |
SVAPAR-94956 |
FS5200, FS7200, FS7300, FS9100, FS9200, FS9500, SVC |
Critical
|
When ISER clustering is configured with a default gateway of 0.0.0.0, the node IPs will not be activated during boot after a reboot or warmstart and the node will remain offline in 550/551 state
(show details)
Symptom |
Loss of Redundancy |
Environment |
ISER clustering |
Trigger |
A node reboot or warmstart |
Workaround |
Manually reconfigure the nodes IP with the 'chnodeip' command |
|
8.6.0.0 |
HyperSwap |
SVAPAR-95349 |
All |
Critical
|
Adding a hyperswap volume copy to a clone of a Volume Group Snapshot may cause all nodes to warmstart, causing a loss of access
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using Volume Group Snapshot and hyperswap |
Trigger |
Creation of a hyperswap volume copy on a clone of a Volume Group Snapshot |
Workaround |
Avoid creation of hyperswap volume copies on clones of a Volume Group Snapshot |
|
8.6.0.0 |
HyperSwap |
HU01782 |
All |
High Importance
|
A node warmstart may occur due to a potentially bad SAS hardware component on the system such as a SAS cable, SAS expander or SAS HIC
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems that have a faulty SAS hardware component |
Trigger |
Faulty SAS component |
Workaround |
None |
|
8.6.0.0 |
Drives |
HU02271 & SVAPAR-88275 |
All |
High Importance
|
A single-node warmstart may occur due to a very low-probability timing window in the thin-provisioning component. This can occur when the partner node has just gone offline, causing a loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with thin-provisioned volumes in standard pools |
Trigger |
Partner node going offline, combined with low-probability timing window |
Workaround |
None |
|
8.6.0.0 |
Thin Provisioning |
HU02339 |
All |
High Importance
|
Multiple node warmstarts can occur if a system has direct Fibre Channel connections to an IBM i host, causing loss of access to data
(show details)
Symptom |
None |
Environment |
Systems running v8.4.0.4 and later, v8.4.2 and later, or v8.5 and later |
Trigger |
Direct attachment to IBM i hosts |
Workaround |
None |
|
8.6.0.0 |
Hosts, Interoperability |
HU02464 |
All |
High Importance
|
An issue in the processing of NVMe host logouts can cause multiple node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems with NVMe hosts |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Hosts, NVMe |
HU02483 |
All |
High Importance
|
T2 Recovery occurred after mkrcrelationship command was run
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running 7.8 or later |
Trigger |
Running the mkrcrelationship command |
Workaround |
None |
|
8.6.0.0 |
Command Line Interface, Global Mirror, Global Mirror With Change Volumes |
HU02488 |
All |
High Importance
|
Remote Copy partnerships disconnect every 15 minutes with error 987301 (Connection to a configured remote cluster has been lost)
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running 8.5.0 in replication partnerships with systems running 8.2.1 or 8.3.0 |
Trigger |
None |
Workaround |
Upgrade partner systems to 8.3.1 or later |
|
8.6.0.0 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02490 |
FS9500 |
High Importance
|
Upon first boot, or subsequent boots of a FS9500 a 1034 error may appear in the event log that states that the CPU PCIe link is degraded
(show details)
Symptom |
None |
Environment |
FAB4 platforms |
Trigger |
Boot up of the system |
Workaround |
This error can be marked as fixed and is not indicative of a hardware fault in the canister |
|
8.6.0.0 |
Reliability Availability Serviceability |
HU02492 |
SVC |
High Importance
|
Configuration backup can fail after upgrade to v8.5. This only occurs on a very small number of systems that have a particular internal cluster state. If a system is running v8.5 and does not have an informational eventlog entry with error ID 988100 (CRON job failed), then it is not affected.
(show details)
Symptom |
Configuration |
Environment |
Systems that have upgraded to v8.5 |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Reliability Availability Serviceability |
HU02497 |
All |
High Importance
|
A system with direct Fibre Channel connections to a host, or to another Spectrum Virtualize system, might experience multiple node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v8.4.0.4 and later, v8.4.2 and later, or v8.5 with direct-attached Fibre Channel connections |
Trigger |
Fibre Channel direct-attached hosts |
Workaround |
Connect affected HBAs via a Fibre Channel switch |
|
8.6.0.0 |
Hosts, Interoperability |
HU02507 |
All |
High Importance
|
A timing window exists in the code that handles host aborts for an ATS (Atomic Test and Set) command, if the host is NVMe-attached. This can cause repeated node warmstarts.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running V8.5.0 or V8.5.1 |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Host Cluster, Hosts |
HU02511 |
All |
High Importance
|
Code version 8.5.0 includes a change in the driver setting for the 25Gb ethernet adapter. This change can cause port errors, which in turn can cause iSCSI path loss symptoms
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running V8.5.0 |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Host Cluster, Hosts, SCSI Unmap, iSCSI |
HU02512 |
FS5000 |
High Importance
|
An FS5000 system with a Fibre Channel direct-attached host can experience multiple node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v8.5 |
Trigger |
Fibre Channel direct-attached hosts |
Workaround |
None |
|
8.6.0.0 |
Hosts |
HU02523 |
All |
High Importance
|
False Host WWPN state shows as degraded for direct attached host after upgrading to 8.5.0.2
(show details)
Symptom |
None |
Environment |
Systems running V8.5.0 |
Trigger |
None |
Workaround |
This does not affect the function or failover behaviour of a host so it can be ignored |
|
8.6.0.0 |
Host Cluster, Hosts, System Update |
HU02525 |
FS5000, FS5100, FS5200, FS7200, FS7300, FS9100, FS9200, FS9500, SVC |
High Importance
|
Code versions 8.4.2.x, 8.5.0.0 - 8.5.0.5 and 8.5.1.0 permitted the use of an iSCSI prefix of 0. However, during an upgrade to 8.5.x, this can prevent all iSCSI hosts from re-establishing iSCSI sessions, thereby causing access loss
(show details)
Symptom |
None |
Environment |
Systems running v8.4.2,v8.5.0,v8.5.1 |
Trigger |
iSCSI prefix of 0 |
Workaround |
Change all hosts with a prefix of 0 before upgrading |
|
8.6.0.0 |
Hosts, iSCSI |
HU02529 |
All |
High Importance
|
A single node warmstart may occur due to a rare timing window, when a disconnection occurs between two systems in an IP replication partnership
(show details)
Symptom |
Single Node Warmstart |
Environment |
None |
Trigger |
None |
Workaround |
Investigate and correct the reason for the network instability between the two sites |
|
8.6.0.0 |
|
HU02530 |
All |
High Importance
|
Upgrades from 8.4.2 or 8.5 fail to start on some platforms
(show details)
Symptom |
None |
Environment |
System is running 8.4.2 or 8.5 with > 1 DRP pools. Seen on DH8 nodes but may affect other types |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
System Update |
HU02534 |
All |
High Importance
|
When upgrading from 7.8.1.5 to 8.5.0.4, PowerHA stops working due to SSH configuration changes
(show details)
Symptom |
None |
Environment |
Systems running 8.5 and higher |
Trigger |
None |
Workaround |
The PowerHA script 'cl_verify_svcpprc_config' can be changed to use the actual username instead of 'admin' |
|
8.6.0.0 |
Reliability Availability Serviceability |
HU02538 |
All |
High Importance
|
Some systems may suffer a thread locking issue caused by the background copy / cleaning progress for flash copy maps
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running V8.2.1+ |
Trigger |
Small timing window when a map leaves a dependency chain |
Workaround |
None needed, recoverable via single node warmstart |
|
8.6.0.0 |
FlashCopy |
HU02539 |
All |
High Importance
|
If an IP address is moved to a different port on a node, the old routing table entries do not get refreshed. Therefore, the IP address maybe inaccessible through the new port
(show details)
Symptom |
None |
Environment |
All |
Trigger |
Moving an IP address to a different port on a node |
Workaround |
Either reboot the node or assign an IP address that has not been used on the node since it was last rebooted |
|
8.6.0.0 |
|
HU02545 |
All |
High Importance
|
When following the 'removing and replacing a faulty node canister' procedure, the satask chbootdrive -replacecanister fails to clear the reported 545 error - instead the replacement reboots into 525 / 522 service state
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Drives, Reliability Availability Serviceability |
HU02549 |
All |
High Importance
|
When upgrading from a lower level, to 8.5 or higher for the first time, an unexpected node warmstart may occur that can lead to a stalled upgrade
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems upgrading from a lower level to 8.5 or higher |
Trigger |
First time upgrade to 8.5 or higher from a lower level |
Workaround |
None |
|
8.6.0.0 |
System Update |
HU02555 |
All |
High Importance
|
A node may warmstart if the system is configured for remote authorization, but no remote authorization service, such as LDAP, has been configured
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running 8.5.x onwards |
Trigger |
Attempted login by a user that is not a locally configured user such as LDAP |
Workaround |
Log in to the node with a known local user (i.e. superuser) then run 'svctask chauthservice -enable no -type ldap' to correct the auth inconsistency |
|
8.6.0.0 |
LDAP |
HU02558 |
FS5000, FS5100, FS5200, FS7200, FS7300, FS9100, FS9200, FS9500, SVC |
High Importance
|
A timing window exists if a node encounters repeated timeouts on I/O compression requests. This can cause two threads to conflict with each other, thereby causing a deadlock condition to occur.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.2 and later, v8.3 and later, v8.4 and later, or v8.5 and later |
Trigger |
None |
Workaround |
None. When the node detects the deadlock condition, it warmstarts in order to clear the issue |
|
8.6.0.0 |
Compression |
HU02562 |
All |
High Importance
|
A node can warmstart when a 32 Gb Fibre Channel adapter receives an unexpected asynchronous event via internal mailbox commands. This is a transient failure caused during DMA operations
(show details)
Symptom |
Loss of Access to Data |
Environment |
Any system with a 32 Gb Fibre Channel adapter installed and running code level 8.4.0.4 or higher |
Trigger |
A 32 Gb Fibre Channel adapter receives an unexpected asynchronous event via internal mailbox commands |
Workaround |
None |
|
8.6.0.0 |
|
HU02565 |
All |
High Importance
|
Node warmstart when generating data compression savings data for 'lsvdiskanalysis'
(show details)
Symptom |
Single Node Warmstart |
Environment |
All |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
|
HU02569 |
All |
High Importance
|
Due to a low-probability timing window, when processing I/O from both SCSI and NVMe hosts, a node may warmstart to clear the condition
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running 8.5.x when processing I/O from both SCSI and NVMe hosts |
Trigger |
Stale NVME command encountered in the control block |
Workaround |
Upgrade to 8.5.3.0 |
|
8.6.0.0 |
Host Cluster, Hosts, NVMe, SCSI Unmap, iSCSI |
HU02573 |
All |
High Importance
|
HBA firmware can cause a port to appear to be flapping. The port will not work again until the HBA is restarted by rebooting the node.
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with Fibre Channel adapters |
Trigger |
Effects systems with high utilization, and possibly bursty IO |
Workaround |
Rebooting the node will reset the buffer, thereby allowing the port to login again |
|
8.6.0.0 |
Fibre Channel, Hosts |
HU02580 |
All |
High Importance
|
If FlashCopy mappings are force stopped, and the targets are in a remote copy relationship, then a node may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running 8.5.2 with FlashCopy mappings that have been force stopped and the targets are in a remote copy relationship |
Trigger |
FlashCopy mappings that have been force stopped and the targets are in a remote copy relationship |
Workaround |
Delete and recreate the Flashcopy mappings, or stop the Flashcopy mappings without using the force option |
|
8.6.0.0 |
FlashCopy |
HU02581 |
All |
High Importance
|
Due to a low probability timing window, a node warmstart might occur when I/O is sent to a partner node and before the partner node recognizes that the disk is online
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
Failing I/O on the remote node causes the local I/O to fail, believing the path is about to go offline, but it never does, thereby causing I/O to be queued indefinitely, leading to a node warmstart. |
Workaround |
None |
|
8.6.0.0 |
Cache |
HU02583 |
All |
High Importance
|
FCM drive ports maybe excluded after a failed drive firmware download. Depending on the number of drives impacted, this may take the RAID array offline
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.3.1 |
Trigger |
Drive firmware upgrade |
Workaround |
Physical reseat or power cycle the drive |
|
8.6.0.0 |
Drives |
HU02589 |
FS5200, FS7200, FS9100, FS9200, FS9500 |
High Importance
|
Reducing the expiration date of snapshots can cause volume creation and deletion to stall
(show details)
Symptom |
None |
Environment |
Systems running V8.4.2,V8.5.0,V8.5.1,V8.5.2 |
Trigger |
Reducing the expiration date of snapshots |
Workaround |
Before the oldest snapshot of the new policy is about to be deleted, manually delete some of the older snapshots from the previous policy to prevent deletion overlap |
|
8.6.0.0 |
FlashCopy, Policy-based Replication, Safeguarded Copy & Safeguarded Snapshots |
IT41191 |
All |
High Importance
|
If a REST API client authenticates as an LDAP user, a node warmstart can occur
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.5 |
Trigger |
REST API authentication with LDAP user |
Workaround |
Use a locally-authenticated user instead of an LDAP user |
|
8.6.0.0 |
REST API |
IT41835 |
All |
High Importance
|
A T2 recovery may occur when a failed drive in the system is replaced with an unsupported drive type
(show details)
Symptom |
Loss of Access to Data |
Environment |
System that have drives reporting as UNSUPPORTED |
Trigger |
A drive with a tech type as UNSUPPORTED will cause this T2 during drive replacement. |
Workaround |
The system should recover automatically. To prevent the issue in the future, make sure system supported drive is used during replacement |
|
8.6.0.0 |
Drives |
SVAPAR-82950 |
FS9500, SVC |
High Importance
|
If a FlashSystem 9500 or SV3 node had a USB Flash Drive present at boot, upgrading to either 8.5.0.7 or 8.5.3.0 may cause the node to become unresponsive. Systems already running 8.5.0.7 or 8.5.3.0 are not affected by this issue
(show details)
Symptom |
Loss of Redundancy |
Environment |
FlashSystem 9500 or SV3 node with a USB Flash Drive present |
Trigger |
Upgrade to 8.5.3.0 |
Workaround |
None |
|
8.6.0.0 |
Reliability Availability Serviceability |
SVAPAR-83290 |
FS5000 |
High Importance
|
An issue with the Trusted Platform Module (TPM) in FlashSystem 50xx nodes may cause the TPM to become unresponsive. This can happen after a number of weeks of continuous runtime.
(show details)
Symptom |
Single Node Warmstart |
Environment |
FS50xx platforms running either V8.4.0,V8.4.1,V8.4.2,V8.5.0,V8.5.1,V8.5.2,V8.5.3 |
Trigger |
Unresponsive TPM |
Workaround |
Reboot each node in turn. Wait 30 minutes between the two nodes in an I/O group, to allow hosts to failover. Check there are no volumes dependent on the second node before proceeding with the reboot. After all nodes have been rebooted, retry the configuration action, which should now complete successfully. |
|
8.6.0.0 |
|
SVAPAR-84305 |
All |
High Importance
|
A node may warmstart when attempting to run 'chsnmpserver -community' command without any additional parameter
(show details)
Symptom |
Loss of Access to Data |
Environment |
Any platform running V8.4.0 |
Trigger |
None |
Workaround |
Use an additional parameter with the 'chsnmpserver -community' command |
|
8.6.0.0 |
System Monitoring |
SVAPAR-84331 |
All |
High Importance
|
A node may warmstart when the 'lsnvmefabric -remotenqn' command is run
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system running NVMe |
Trigger |
The warmstart can occur typically when the 'lsnvmefabric -remotenqn' command is run by a script or orchestration layer such as Redhat Openshift or Kubernetes, combined with the IBM CSI driver. |
Workaround |
None |
|
8.6.0.0 |
NVMe |
SVAPAR-85093 |
All |
High Importance
|
Systems that are using Policy-Based Replication may experience node warmstarts, if host I/O consists of large write I/Os with a high queue depth
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems with Policy-Based Replication and heavy large write I/O workloads |
Trigger |
Systems configured to use Policy-Based Replication and have host I/O consisting of large write I/Os that have a high queue depth |
Workaround |
None |
|
8.6.0.0 |
Policy-based Replication |
SVAPAR-85396 |
FS5000, FS5100, FS5200, FS7200, FS7300, FS9100, FS9200, FS9500 |
High Importance
|
Replacement Samsung NVME drives may show as unsupported, or they may fail during a firmware upgrade as unsupported, due to a VPD read problem
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running V8.5.3 with industry standard NVMe drives |
Trigger |
Drive firmware update or drive replacements |
Workaround |
Manually power cycling the slot of the failed drive often helps |
|
8.6.0.0 |
Drives |
SVAPAR-86035 |
All |
High Importance
|
Whilst completing a request, a DRP pool attempts to allocate additional metadata space, but there is no free space available. This causes the node to warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any DRP pool that has run out of metadata space |
Trigger |
Not enough metadata space available |
Workaround |
Add additional space to the DRP pool |
|
8.6.0.0 |
Data Reduction Pools |
SVAPAR-89780 |
All |
High Importance
|
A node may warmstart after running the flashcopy command 'stopfcconsistgrp' due to the flashcopy maps in the consistency group being in an invalid state
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system running Flashcopy |
Trigger |
Running the 'stopfcconsistgrp' command, when the flashcopy maps are in an invalid state |
Workaround |
None |
|
8.6.0.0 |
FlashCopy |
SVAPAR-89951 |
All |
High Importance
|
A single node warmstart might occur when a volume group with a replication policy switches the replication to cycling mode.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with policy-based replication |
Trigger |
Replication is switching from journaling to cycling mode |
Workaround |
Contact IBM support for an action plan to force replication into cycling mode permanently |
|
8.6.0.0 |
Policy-based Replication |
SVAPAR-90395 |
FS9500, SVC |
High Importance
|
FS9500 and SV3 might suffer from poor Remote Copy performance due to a lack of internal messaging resources
(show details)
Symptom |
Performance |
Environment |
FS9500 or SV3 systems with Remote Copy, typically running Hyperswap, Metro Mirror, Global Mirror and GMCV |
Trigger |
Not enough resources available for Remote Copy |
Workaround |
None |
|
8.6.0.0 |
Global Mirror, Global Mirror With Change Volumes, HyperSwap, Metro Mirror |
SVAPAR-92066 |
All |
High Importance
|
Node warmstarts can occur after running the 'lsvdiskfcmapcopies' command if Safeguarded Copy is used
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Any 8.5.x system that uses Safeguarded copy |
Trigger |
Running the 'lsvdiskfcmapcopies' command when there is a large number of Flash Copies (256) configured |
Workaround |
None |
|
8.6.0.0 |
Safeguarded Copy & Safeguarded Snapshots |
SVAPAR-92983 |
All |
High Importance
|
There is an issue that prevents Remote users with SSH key to connect to the storage system if BatchMode is enabled
(show details)
Symptom |
Configuration |
Environment |
Any system running 8.5 or later |
Trigger |
Remote users with SSH Key trying to connect to the system using BatchMode |
Workaround |
None |
|
8.6.0.0 |
Security |
SVAPAR-93054 |
All |
High Importance
|
Backend systems on 8.2.1 and beyond have an issue that causes capacity information updates to stop after a T2 or T3 is performed. This affects all backend systems with FCM arrays
(show details)
Symptom |
Configuration |
Environment |
Backend system with FCM arrays on 8.2.1 and beyond |
Trigger |
Systems that have performed either a T2 or T3 recovery |
Workaround |
Upgrade the backend system as the upgrade restarts the capacity update process |
|
8.6.0.0 |
Backend Storage |
SVAPAR-93309 |
All |
High Importance
|
A node may briefly go offline after a battery firmware update
(show details)
Symptom |
Single Node Warmstart |
Environment |
All Storage Virtualize based systems |
Trigger |
Battery firmware update |
Workaround |
None |
|
8.6.0.0 |
System Update |
SVAPAR-94686 |
All |
High Importance
|
The GUI can become slow and unresponsive due to a steady stream of configuration updates such as 'svcinfo' queries for the latest configuration data
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Graphical User Interface |
SVAPAR-99273 |
All |
High Importance
|
If a SAN switch's Fabric Controller issues an abort (ABTS) command, and then issues an RSCN command before the abort has completed, this unexpected switch behaviour can trigger a single-node warmstart.
(show details)
Symptom |
Single Node Warmstart |
Environment |
None |
Trigger |
An unexpected sequence of Fibre Channel frames received from Fabric Controller |
Workaround |
None |
|
8.6.0.0 |
|
HU02446 |
All |
Suggested
|
An invalid alert relating to GMCV freeze time can be displayed
(show details)
Symptom |
None |
Environment |
Systems running 8.4 or later using GMCV |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Global Mirror With Change Volumes |
HU02453 |
All |
Suggested
|
It may not be possible to connect to GUI or CLI without a restart of the Tomcat server
(show details)
Symptom |
None |
Environment |
Systems running v8.4 or later |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Command Line Interface, Graphical User Interface |
HU02463 |
All |
Suggested
|
LDAP user accounts can become locked out because of multiple failed login attempts
(show details)
Symptom |
None |
Environment |
Systems where LDAP accounts use one-time passwords |
Trigger |
None |
Workaround |
Use one or more of the following options: Use CLI instead of GUI. With CLI, after auth cache expires, do not issue any more commands after experiencing a CMMVC7069E error. Log out and log back in with new LDAP password; Max out authentication cache - "chldap -authcacheminutes <min>". Maximum value is 1440min = 24h. Setting it to 600 minutes can minimize probability of hitting the issue; Disable account lock-out on LDAP. |
|
8.6.0.0 |
Graphical User Interface, LDAP |
HU02468 |
All |
Suggested
|
lsvdisk preferred_node_id filter not working correctly
(show details)
Symptom |
None |
Environment |
Systems running 8.4.2 or later |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Command Line Interface |
HU02484 |
All |
Suggested
|
The GUI does not allow expansion of DRP thin or compressed volumes
(show details)
Symptom |
None |
Environment |
Systems running 8.5 |
Trigger |
None |
Workaround |
Use the expandvdisksize CLI command instead |
|
8.6.0.0 |
Data Reduction Pools, Graphical User Interface |
HU02487 |
All |
Suggested
|
Problems expanding the size of a volume using the GUI
(show details)
Symptom |
None |
Environment |
Systems running 8.2.1 or later |
Trigger |
None |
Workaround |
Use the equivalent CMD line command instead such as expandvdisksize |
|
8.6.0.0 |
Graphical User Interface |
HU02491 |
All |
Suggested
|
On upgrade from v8.3.x, v8.4.0 or v8.4.1 to v8.5, if the system has Global Mirror with Change Volumes relationships, a single node warmstart can occur
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems upgrading to v8.5 with GMCV |
Trigger |
None |
Workaround |
Stop all GMCV relationships before upgrading |
|
8.6.0.0 |
Global Mirror With Change Volumes |
HU02494 |
All |
Suggested
|
A system with a DNS server configured, which cannot ping the server, will log information events in the eventlog. In some environments the firewall blocks ping packets but allows DNS lookup, so this APAR disables these events.
(show details)
Symptom |
Configuration |
Environment |
Systems running v8.4.1, v8.4.2 or v8.5 with DNS servers that cannot be pinged. |
Trigger |
Firewall rules that block ping to DNS server |
Workaround |
Change firewall configuration to allow ping to DNS server |
|
8.6.0.0 |
Reliability Availability Serviceability |
HU02498 |
All |
Suggested
|
If a host object with no ports exists on upgrade to v8.5, the GUI volume mapping panel may fail to load.
(show details)
Symptom |
Configuration |
Environment |
Systems running v8.5 |
Trigger |
Upgrade to v8.5, when a host object has no ports specified |
Workaround |
None |
|
8.6.0.0 |
Graphical User Interface |
HU02501 |
All |
Suggested
|
If an internal I/O timeout occurs in a RAID array, a node warmstart can occur
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.5 |
Trigger |
I/O timeout on a RAID array |
Workaround |
None |
|
8.6.0.0 |
RAID |
HU02503 |
All |
Suggested
|
The Date / Time panel can fail to load in the GUI when a timezone set via the CLI is not supported by the GUI
(show details)
Symptom |
None |
Environment |
Systems running 8.5 or later |
Trigger |
Use CLI to set a timezone that is not supported in the GUI |
Workaround |
Configure the time zone via CLI |
|
8.6.0.0 |
Graphical User Interface |
HU02504 |
All |
Suggested
|
The Date / Time panel can display an incorrect timezone and default to manual time setting rather than NTP
(show details)
Symptom |
None |
Environment |
Systems running 8.5 or later |
Trigger |
None |
Workaround |
Configure the time zone via CLI |
|
8.6.0.0 |
Graphical User Interface |
HU02505 |
All |
Suggested
|
A single node warmstart can occur on v8.5 systems running DRP, due to a low-probability timing window during normal running
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.5 |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Data Reduction Pools |
HU02508 |
All |
Suggested
|
The mkippartnership cli command does not allow a portset with a space in the name as a parameter.
(show details)
Symptom |
None |
Environment |
Systems running V8.5.0 |
Trigger |
portset containing a space |
Workaround |
Request customer recreate portset name without a space |
|
8.6.0.0 |
Command Line Interface |
HU02509 |
All |
Suggested
|
Upgrade to v8.5 can cause a single node warmstart, if nodes previously underwent a memory upgrade while DRP was in use
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.5 |
Trigger |
Upgrade to v8.5, after memory upgrade was performed while DRP was in use. |
Workaround |
None |
|
8.6.0.0 |
Data Reduction Pools |
HU02514 |
All |
Suggested
|
Firmware upgrade may fail for certain drive types, with the error message CMMVC6567E The Apply Drive Software task cannot be initiated because no download images were found in the package file
(show details)
Symptom |
Configuration |
Environment |
Systems running v8.5 |
Trigger |
Drive firmware upgrade |
Workaround |
None |
|
8.6.0.0 |
Drives |
HU02515 |
FS9500 |
Suggested
|
Fan speed on FlashSystem 9500 can be higher than expected, if a high drive temperature is detected
(show details)
Symptom |
None |
Environment |
FlashSystem 9500 running v8.5 |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Drives |
HU02528 |
All |
Suggested
|
When upgrading to 8.5.0 or higher, a situation may occur whereby a variable is not locked at the correct point, resulting in a mismatch. The system code detects this and initiates a warmstart to reset any erroneous values
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems that have upgraded to v8.5 |
Trigger |
Can only occur during upgrade to 8.5 or higher |
Workaround |
None |
|
8.6.0.0 |
Reliability Availability Serviceability |
HU02544 |
FS5200, FS7200, FS7300, FS9100, FS9200, FS9500, SVC |
Suggested
|
On systems running 8.5.2.1, if you are not logged in as superuser and you try to create a partnership for policy-based replication, or enable policy-based replication on an existing partnership, then this can trigger a single node warmstart.
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system running 8.5.2.1, with Policy-based replication configured |
Trigger |
Creating a partnership for Policy-based replication when not logged in as superuser |
Workaround |
Retry the operation as superuser |
|
8.6.0.0 |
Policy-based Replication |
HU02553 |
FS9500, SVC |
Suggested
|
Remote copy relationships may not correctly display the name of the vdisk on the remote cluster
(show details)
Symptom |
None |
Environment |
Any FS9500 or SV3 node running 8.4.2 or later, and using remote copy |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02559 |
All |
Suggested
|
A GUI resource issue may cause an out-of-memory condition, leading to the CIMOM and GUI becoming unresponsive, or showing incomplete information
(show details)
Symptom |
None |
Environment |
Systems running v8.3.1,v8.4.0,v8.4.1,v8.4.2,v8.5.0 |
Trigger |
None |
Workaround |
Restart the cimserver service or tomcat service |
|
8.6.0.0 |
Graphical User Interface |
HU02568 |
All |
Suggested
|
Unable to create remote copy relationship with 'mkrcrelationship' with Aux volume ID greater than 10,000 when one of the systems in the set of partnered systems is limited to 10,000 volumes, either due to the limits of the platform (hardware) or the installed software version
(show details)
Symptom |
None |
Environment |
Any V8.5.x system with Aux volume ID greater than 10,000 |
Trigger |
If either of the set of partnered systems is limited to 10,000 volumes |
Workaround |
Use vdisk IDs for Remote Copy across all systems that are within the limits of the lowest partnered system |
|
8.6.0.0 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02579 |
All |
Suggested
|
The GUI 'Add External iSCSI Storage' wizard does not work with portsets. The ports are shown but are not selectable
(show details)
Symptom |
None |
Environment |
Any system running V8.4.2,V8.5.0,V8.5.1,V8.5.2, and using IP portsets |
Trigger |
None |
Workaround |
Use the command line to configure the external iSCSI connection |
|
8.6.0.0 |
Graphical User Interface, iSCSI |
HU02592 |
All |
Suggested
|
In some scenarios DRP can request RAID to attempt a read by reconstructing data from other strips. In certain cases this can result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Any system running 8.5.x with DRP configured |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Data Reduction Pools, RAID |
HU02594 |
All |
Suggested
|
Initiating drive firmware update via management user interface for one drive class can prompt all drives to be updated
(show details)
Symptom |
None |
Environment |
Any system running V8.4.2,V8.5.0,V8.5.1,V8.5.2,V8.5.3 |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Drives, System Update |
HU02600 |
All |
Suggested
|
Single node warmstart caused by a rare race condition triggered by multiple aborts and I/O issues
(show details)
Symptom |
None |
Environment |
None specific |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
|
SVAPAR-84099 |
All |
Suggested
|
An NVMe codepath exists where by strict state checking incorrectly decides that a software flag state is invalid, there by triggering a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with NVMe hosts |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
Hosts, NVMe |
SVAPAR-85640 |
FS5000, FS5100, FS5200, FS7200, FS7300, FS9100, FS9200, FS9500 |
Suggested
|
If new nodes/iogroups are added to an SVC cluster that is virtualizing a clustered SpecV system, an attempt to add the SVC node host objects to a host cluster on the backend SpecV system will fail with CLI error code CMMVC8278E due to incorrect policing
(show details)
Symptom |
Configuration |
Environment |
Clustered SpecV system virtualized behind an SVC cluster or other SpecV cluster |
Trigger |
Running the 'addhostclustermember' command |
Workaround |
None |
|
8.6.0.0 |
Host Cluster |
SVAPAR-86182 |
All |
Suggested
|
A node may warmstart if there is an encryption key error that prevents a distributed raid array from being created
(show details)
Symptom |
Single Node Warmstart |
Environment |
All enclosure based systems using encryption |
Trigger |
Creating a distributed raid array |
Workaround |
None |
|
8.6.0.0 |
Distributed RAID, Encryption |
SVAPAR-89296 |
All |
Suggested
|
Immediately after upgrade from pre-8.4.0 to 8.4.0 or later, EasyTier may stop promoting hot data to the tier0_flash tier if it contains non-FCM storage. This issue will automatically resolve on the next upgrade
(show details)
Symptom |
Performance |
Environment |
Multi-tier pools where tier0_flash contains non-FCM storage |
Trigger |
Upgrading from pre-8.4.0 to 8.4.0 |
Workaround |
Upgrade to any later version of software, or warmstart the config node |
|
8.6.0.0 |
EasyTier |
SVAPAR-89781 |
All |
Suggested
|
The 'lsportstats' command does not work via the REST API until code level 8.5.4.0
(show details)
Symptom |
Configuration |
Environment |
Any system running 8.4.0 or higher |
Trigger |
None |
Workaround |
None |
|
8.6.0.0 |
|
SVAPAR-93442 |
All |
Suggested
|
User ID does not have the authority to submit a command in some LDAP environments
(show details)
Symptom |
None |
Environment |
A Remote user ID is defined to multiple groups within a directory service, one of which is a custom defined group |
Trigger |
Configure storage cluster for remote authentication. Define a given ID in multiple access groups, one of which is a customized, with varying access levels |
Workaround |
To avoid this issue the currently defined user groups need to be deleted and then re-created in a different order, ensuring the group with the higher access authority is defined first. |
|
8.6.0.0 |
|
SVAPAR-93987 |
All |
Suggested
|
A timeout may cause a single node warmstart, if a FlashCopy configuration change occurs while there are many I/O requests outstanding for a source volume which has multiple FlashCopy targets
(show details)
Symptom |
Single Node Warmstart |
Environment |
FlashCopy with multiple targets for a single source volume |
Trigger |
Timing window during FlashCopy configuration change |
Workaround |
None |
|
8.6.0.0 |
FlashCopy |
SVAPAR-94682 |
All |
Suggested
|
SMTP fails if the length of the email server's domain name is longer than 40 characters
(show details)
Symptom |
Configuration |
Environment |
Having an email server which has a domain name longer than 40 characters |
Trigger |
None |
Workaround |
Use the IP address or shorter domain name < 40 characters for the email server |
|
8.6.0.0 |
|
SVAPAR-94703 |
All |
Suggested
|
The estimated compression savings value shown in the GUI for a single volume is incorrect. The total savings for all volumes in the system will be shown
(show details)
Symptom |
Configuration |
Environment |
None |
Trigger |
None |
Workaround |
Use the CLI 'lsvdiskanalysis' command |
|
8.6.0.0 |
Graphical User Interface |
SVAPAR-94902 |
FS5000, FS5100, FS5200, FS7200, FS7300, FS9100, FS9200, FS9500 |
Suggested
|
When attempting to enable local port masking for a specific subset of control enclosure based clusters, this may fail with the following message; 'The specified port mask cannot be applied because insufficient paths would exist for node communication'
(show details)
Symptom |
None |
Environment |
Direct attached control enclosures and individual nodes that do not have fibre channel paths to their partner node within the same enclosure |
Trigger |
Attempting to enable local port masking |
Workaround |
Enable local port masking for direct attached |
|
8.6.0.0 |
|
SVAPAR-96656 |
All |
Suggested
|
VMware hosts may experience errors creating snapshots, due to an issue in the VASA Provider
(show details)
Symptom |
Configuration |
Environment |
Any system leveraging VASA Provider functionality |
Trigger |
VASA Provider related error |
Workaround |
Applicable workaround dependent on context of issue |
|
8.6.0.0 |
|