HU01720 |
All |
HIPER
|
An issue in the handling of compressed volume shrink operations, in the presence of EasyTier migrations, can cause DRAID MDisk timeouts leading to an offline MDisk group
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1 or later using EasyTier with compressed volumes |
Trigger |
None |
Workaround |
None |
|
8.1.1.2 |
Compression, EasyTier |
HU01792 |
All |
HIPER
|
When a DRAID array has multiple drive failures and the number of failed drives is greater than the number of rebuild areas in the array it is possible that the storage pool will be taken offline during the copyback phase of a rebuild. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
8.1.1.2 |
Distributed RAID |
HU01767 |
All |
Critical
|
Reads of 4K/8K from an array can under exceptional circumstances return invalid data. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8.0 or earlier |
Trigger |
None |
Workaround |
None |
|
8.1.1.2 |
RAID, Thin Provisioning |
HU01769 |
All |
Critical
|
Systems with DRAID arrays, with more than 131,072 extents, may experience multiple warmstarts due to a backend SCSI UNMAP issue
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1.1 or later |
Trigger |
Create a DRAID array with >131,072 extents on SSDs |
Workaround |
Disable UNMAP at a system level by issuing a "svctask chsystem -unmap off" command |
|
8.1.1.2 |
Distributed RAID |
HU01771 |
SVC, V7000 |
High Importance
|
An issue with the CMOS battery in a node can cause an unexpectedly large log file to be generated by the BMC. At log collection the node may be taken offline
(show details)
Symptom |
Loss of Redundancy |
Environment |
SVC & V7000 systems running v7.8 or later |
Trigger |
Node CMOS battery issue |
Workaround |
None |
|
8.1.1.2 |
System Monitoring |
HU01619 |
All |
Suggested
|
A misreading of the PSU register can lead to failure events being logged incorrectly
(show details)
Symptom |
None |
Environment |
Systems running v7.6 or later |
Trigger |
None |
Workaround |
None |
|
8.1.1.2 |
System Monitoring |
HU01664 |
All |
Suggested
|
A timing window issue during an upgrade can cause the node restarting to warmstart stalling the upgrade
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
8.1.1.2 |
System Update |
HU01740 |
All |
Suggested
|
The timeout setting for key server commands may be too brief, when the server is busy, causing those commands to fail
(show details)
Symptom |
None |
Environment |
Systems running v7.8 or later using encryption |
Trigger |
Enter mkkeyserver command |
Workaround |
Retry command |
|
8.1.1.2 |
Encryption |
HU01756 |
V7000 |
Suggested
|
A scheduling issue may cause a config node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize V7000 Gen 2 systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
8.1.1.2 |
|
HU00247 |
All |
Critical
|
A rare deadlock condition can lead to a RAID5 or RAID6 array rebuild stalling at 99%
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using RAID5/6 |
Trigger |
None |
Workaround |
None |
|
8.1.1.1 |
Distributed RAID, RAID |
IT19192 |
All |
Critical
|
An issue in the handling of GUI certificates may cause warmstarts leading to a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.1.1.1 |
Graphical User Interface, Reliability Availability Serviceability |
IT23747 |
All |
High Importance
|
For large drive sizes the DRAID rebuild process can consume significant CPU resource adversely impacting system performance
(show details)
Symptom |
Performance |
Environment |
Systems running v7.6 or later using DRAID |
Trigger |
High write workloads during a DRAID rebuild |
Workaround |
Reduce write workload during rebuild or use a volume/pool level throttle |
|
8.1.1.1 |
Distributed RAID |
HU01655 |
All |
Suggested
|
The algorithm used to calculate an SSDs replacement date can sometimes produce incorrect results leading to a premature End-of-Life error being reported
(show details)
Symptom |
None |
Environment |
Systems using SSDs |
Trigger |
None |
Workaround |
Ignore error if SSD write endurance (in lsdrive CLI output) shows less than 80 |
|
8.1.1.1 |
Drives |
HU01730 |
SVC |
Suggested
|
When running the DMP for a 1046 error the picture may not indicate the correct position of the failed adapter
(show details)
Symptom |
None |
Environment |
DH8 systems |
Trigger |
Run DMP for 1046 error |
Workaround |
Be aware of which adapter is failed and do not use the picture in the GUI |
|
8.1.1.1 |
GUI Fix Procedure |
HU01763 |
SVC |
Suggested
|
A single node warmstart may occur on a DH8 config node when inventory email is created. The issue only occurs if this coincides with a very high rate of CLI commands and high I/O workload on the config node
(show details)
Symptom |
Single Node Warmstart |
Environment |
DH8 systems running v7.7.1 or later |
Trigger |
High CLI workload and I/O workload at same time as inventory email |
Workaround |
None |
|
8.1.1.1 |
Command Line Interface, System Monitoring |
HU01726 |
All |
HIPER
|
A slow raid member drive in an MDisk may cause node warmstarts and the MDisk to go offline for a short time
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using DRAID |
Trigger |
Failing drives |
Workaround |
None |
|
8.1.1.0 |
Distributed RAID |
HU01618 |
All |
Critical
|
When using the charraymember CLI command if a member id is entered that is greater than the maximum number of members in a TRAID array then a T2 recovery will be initiated
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Enter a charraymember command with a member id greater than the maximum value for the related RAID type |
Workaround |
Ensure member id is less than or equal to maximum number of possible members in the array |
|
8.1.1.0 |
RAID |
HU01620 |
All |
Critical
|
Configuration changes can slow critical processes and, if this coincides with cloud account statistical data being adjusted, a Tier 2 recovery may occur
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8 or later that support TCT |
Trigger |
Configuration changes around midnight |
Workaround |
Avoid making large numbers of configuration changes around midnight |
|
8.1.1.0 |
Transparent Cloud Tiering |
HU01671 |
All |
Critical
|
Metadata between two nodes in an I/O group can become out of step leaving one node unaware of work scheduled on its partner. This can lead to stuck array synchronisation and false 1691 events
(show details)
Symptom |
Offline Volumes |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.1.1.0 |
RAID |
HU01678 |
All |
Critical
|
Entering an invalid parameter in the addvdiskaccess command may initiate a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Include an invalid argument in an addvdiskaccess command |
Workaround |
Use care when entering addvdiskaccess CLI commands |
|
8.1.1.0 |
Command Line Interface |
HU01701 |
SVC |
Critical
|
Following loss of all logins to an external controller, that is providing quorum, when the controller next logs in it will not be automatically used for quorum
(show details)
Symptom |
Offline Volumes |
Environment |
SVC systems |
Trigger |
Loss of all logins to a controller providing quorum |
Workaround |
Set allow_quorum with TRUE to all controllers |
|
8.1.1.0 |
HyperSwap |
HU01420 |
All |
High Importance
|
An issue in DRAID can cause repeated node warmstarts in the circumstances of a degraded copyback operation to a drive
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
8.1.1.0 |
Distributed RAID |
HU01525 |
All |
High Importance
|
During an upgrade a resource locking issue in the compression component can cause a node to warmstart multiple times and become unavailable
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.6 or earlier using compressed volumes |
Trigger |
Upgrade to v7.7 or later |
Workaround |
None |
|
8.1.1.0 |
Compression, System Update |
HU01632 |
All |
High Importance
|
A congested fabric causes the Fibre Channel adapter firmware to abort I/O resulting in node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All systems |
Trigger |
Congested fabric |
Workaround |
Avoid fabric congestion |
|
8.1.1.0 |
Reliability Availability Serviceability |
HU01190 |
All |
Suggested
|
Where a controller, which has been assigned to a specific site, has some logins intentionally removed then the system can continue to display the controller as degraded even when the DMP has been followed and errors fixed
(show details)
Symptom |
None |
Environment |
Systems with site ids assigned (e.g. stretched clusters) |
Trigger |
Logins removed |
Workaround |
The degraded issue can be cleared by clearing the event log. |
|
8.1.1.0 |
Backend Storage |
HU01462 |
V7000, V5000 |
Suggested
|
Environmental factors can trigger a protection mechanism, that causes the SAS chip to freeze, resulting in a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize systems |
Trigger |
None |
Workaround |
None |
|
8.1.1.0 |
Drives |
HU01512 |
All |
Suggested
|
During a DRAID MDisk copy-back operation a miscalculation of the remaining work may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
8.1.1.0 |
Distributed RAID |
HU01602 |
All |
Suggested
|
When security scanners send garbage data to SVC/Storwize iSCSI target addresses a node warmstart may occur
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
Security scanning SVC/Storwize iSCSI target addresses with random data |
Workaround |
If security probes are required ensure they are not issued within 30 minutes of each other to both nodes in an I/O group |
|
8.1.1.0 |
iSCSI |
HU01633 |
All |
Suggested
|
Even though synchronisation has completed a RAID array may still show progress to be at 99%
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.1.1.0 |
RAID |
HU01654 |
SVC, V7000, V5000 |
Suggested
|
There may be a node warmstart when a switch of direction, in a HyperSwap relationship, fails to complete properly
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later using HyperSwap |
Trigger |
None |
Workaround |
None |
|
8.1.1.0 |
HyperSwap |
HU01688 |
SVC |
Suggested
|
Unexpected used_virtualization figure in lslicense output after upgrade
(show details)
Symptom |
None |
Environment |
SVC systems running v7.8.1 or earlier with A9000 backend controllers |
Trigger |
Upgrade to v8.1 or later with the amount of mapped logical capacity, from an A9000 controller, being greater than the amount of mapped physical capacity |
Workaround |
Use chmdisk to ensure correct tiering |
|
8.1.1.0 |
Command Line Interface |
HU01698 |
All |
Suggested
|
A node warmstart may occur when deleting a compressed volume if a host has written to the volume minutes before the volume is deleted
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with compressed volumes |
Trigger |
Write I/O received by compression code for a volume that has been deleted |
Workaround |
Remove volume host mappings and leave an interval of time (90mins) before removing the volume |
|
8.1.1.0 |
Compression |
HU01747 |
All |
Suggested
|
The incorrect detection of a cache issue can lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
8.1.1.0 |
Cache |
IT20586 |
All |
Suggested
|
Due to an issue in Lancer G5 firmware after a node reboot the LED of the 10GBE port may remain amber even though the port is working normally
(show details)
Symptom |
None |
Environment |
Systems with Lancer G5 HBAs |
Trigger |
None |
Workaround |
None |
|
8.1.1.0 |
Reliability Availability Serviceability |