This document describes the authorized program analysis reports (APARs) resolved in IBM Spectrum Scale 5.1.0.x releases.
This document was last updated on 11th March, 2021.
Tips:
APAR |
Severity |
Description |
Resolved in |
Feature Tags |
---|---|---|---|---|
|
| |||
IJ30552 | High Importance | Migration within the same group pool does not fully rebalance unless higher preferred pools have more available space than less preferred pools. (show details) | 5.1.0.3 | Core GPFS |
IJ30553 | Suggested | The administrator is unable to change the page pool setting on the GNR recovery group server. The problem is seen only on recovery groups not managed by mmvdisk. The mmchconfig command will fail, and the following error message is displayed: The --force-rg-server flag must be used to change the pagepool (show details) |
5.1.0.3 | GNR, ESS |
IJ30641 | Critical | logAssertFailed (*respPP != __null) cacheops.C (show details) | 5.1.0.3 | AFM |
IJ30682 | High Importance | After dm_punch_hole call, a dm_get_allocinfo could return improper results for the information of the data blocks allocation. (show details) | 5.1.0.3 | DMAPI |
IJ30684 | Suggested | After Node-B successfully reestablishes a broken connection to Node-A, Node-A still shows the reconnect_start state (DEGRADED). (show details) | 5.1.0.3 | System health |
IJ30710 | High Importance | When mmbackup or tsapolicy is called to scan files, it could report "no such file or directory" for existing files. (show details) | 5.1.0.3 | mmbackup, DMAPI, AFM, tsapolicy, GPFS API |
IJ30606 | High Importance | An RPC message could be handled twice when TCP reconnect happens This could cause log assertion, FS struct error or be silently handled depending on the type of RPC. (show details) | 5.1.0.3 | Core GPFS |
IJ30681 | Suggested | Hit error 2 while replicating to COS. (show details) | 5.1.0.3 | AFM |
IJ30685 | High Importance | SetAttr operation on renamed object can get requeued forever during AFM replication to the COS backend. (show details) | 5.1.0.3 | AFM |
IJ30712 | Suggested | The CES network status is shown as UNKNOWN on CES nodes if the policy is set to 'node-affinity' and the node is a member of a group. (show details) | 5.1.0.3 | System health |
IJ30777 | Suggested | If the mmvdisk command takes more than 60 seconds to complete, mmhealth reports all pdisks as vanished. On larger systems with many I/O nodes and pdisks, 60 second timeouts are not enough. (show details) | 5.1.0.3 | System health, GUI |
IJ30793 | Suggested | mmcrfs fails to create file systems when the cluster is configured with minQuorumNodes greater than one and tiebreakerDisks are in use. (show details) | 5.1.0.3 | Admin commands |
IJ30757 | Suggested | Upload fails with error 235 when the max-upload-parts option is set to a value higher then what an int data type can hold. (show details) | 5.1.0.3 | AFM |
IJ30795 | Suggested | Running "mmces events list" stdout prints many trailing white characters (empty spaces) unnecessarily. (show details) | 5.1.0.3 | CES |
IJ30786 | HIPER | Revalidation on a AFM fileset fails on a RHEL 8.3 gateway node and home changes may not be detected causing the data or metadata mismatch between cache and home. (show details) | 5.1.0.3 | AFM, AFM DR |
IJ30788 | Critical | With async refresh enabled, file system quiesce is blocked during the remote operation and it might result in a deadlock if the remote is not responding. (show details) | 5.1.0.3 | AFM |
IJ30776 | Critical | Generating hard link remove operations fails during the recovery with error 22. (show details) | 5.1.0.3 | AFM |
IJ30822 | Critical | Daemon (AFM) assert goes off: getReqP->r_length <= ksP->r_bufSize (show details) | 5.1.0.3 | AFM |
IJ30789 | High Importance | GPFS daemon assert going off: "verbsDtoThread_i: fatal device error" in file verbsSendRecv.C or assert "verbsAsyncThread_i: fatal device error" in file verbsInit.C, resulting in a GPFS hard shutdown. (show details) | 5.1.0.3 | RDMA |
IJ30826 | High Importance | mmbackup repeats the backup/expire of the same file after file is changed. (show details) | 5.1.0.3 | mmbackup |
IJ31043 | High Importance | In a mixed AIX/Linux cluster, the mmbackup command could fail with gskit/ssl errors after upgrading IBM Spectrum Protect code to 8.1.11, which introduced new rpm for gskit 8.0-55.17 that is not compatible with gpfs.gskit version. (show details) | 5.1.0.3 | mmbackup |
IJ31044 | HIPER | AFM gateway node crashes if the home is not responding and multiple threads are trying to read the same file. (show details) | 5.1.0.3 | AFM |
IJ30985 | HIPER | AFM gateway node asserts if the home is not responding and multiple threads are trying to read the same file. (show details) | 5.1.0.3 | AFM |
IJ30987 | High Importance | When snapshot is being used, a modification to a file could result in the inode getting copied to the snapshot. If missing, a subsequent inode copy would trigger this assert. (show details) | 5.1.0.3 | Snapshots |
IJ31060 | High Importance | State of Physical disk shown as "unknown" in mmhealth and GUI for ECE. (show details) | 5.1.0.3 | System health, GUI |
IJ31062 | High Importance | When aioSyncDelay config is enabled, the buffer steal and the aio writes that need to be done as buffered I/O may race with each other and causes log assert isSGPanicked in clearBuffer. (show details) | 5.1.0.3 | Core GPFS |
IJ30994 | High Importance | When an RDMA network port for a cluster node dies, all the RDMA connections from all the RDMA adapters are disconnected. The following IBM Spectrum Scale log messages are indicative of this issue: [W] VERBS RDMA async event IBV_EVENT_PORT_ERR on mlx5_2 port 1. [W] VERBS RDMA pkey index for pkey 0xffff changed from -1 to 0 for device mlx5_2 port 1. [W] VERBS RDMA port state changed from IBV_PORT_ACTIVE to IBV_PORT_DOWN for device mlx5_2 port 1 fabnum 0. [I] VERBS RDMA closed connection to 192.168.12.101 (ems5k) on mlx5_0 port 1 fabnum 0 index 0 cookie 1 due to IBV_EVENT_PORT_ERR [I] VERBS RDMA closed connection to 192.168.12.13 (ece-13 in esscluster.mmfsd.net) on mlx5_2 port 1 fabnum 0 index 11 cookie 12 due to IBV_EVENT_PORT_ERR (show details) |
5.1.0.3 | RDMA |
IJ31071 | Suggested | GPFS command reports incorrect default for nsdRAIDMaxRecoveryRetries. (show details) | 5.1.0.3 | Admin commands |
IJ31021 | High Importance | Deleted inodes with inconsistencies were ignored in IBM Spectrum Scale versions prior to 5.1 but are flagged as a corruption in versions 5.1 and later. If such corruptions exist, then they can cause commands such as 'mmchdisk start' to fail. (show details) | 5.1.0.3 | Core GPFS |
IJ31111 | High Importance | mmbackup does not honor --max-backup-size in snapshot backup. (show details) | 5.1.0.3 | mmbackup |
IJ29859 | High Importance | When mmdelnode is issued against a node whose mmfsd daemon is still up, several of the nodes in the cluster can fail with messages like the following: [E] Deleted node 169.28.113.36 (nodeX) is still up. [E] Node 169.28.113.36 (nodeX) has been deleted from the cluster (show details) |
5.1.0.2 | Cluster membership |
IJ29812 | Suggested | In IBM Spectrum Scale Erasure Code Edition, it is possible for all of the server's pdisks (physical disks) to become missing, either due to network failure, node failure, or through a planned "node suspend" maintenance procedure. When this
happens, the system will continue to function if there is sufficient remaining fault tolerance. However, smaller configurations with less ECE nodes are exposed to a race condition where pdisk state changes can interrupt a system-wide descriptor update which causes the recovery group to resign. It is also possible to experience this problem with higher probability when using small ESS configurations, such as the GS1 or GS2 enclosures. For both ESS and ECE, a possible symptom may appear in the mmfs.log in this form when a pdisk state change is quickly followed by a resign message claiming VCD write failures before the system fault tolerance is exceeded: 2020-12-01_19:01:36.696-0400: [D] Pdisk n004p005 of RG rg1 state changed from ok/00000.180 to missing/ suspended/00050.180. 2020-12-01_19:01:36.697-0400: [E] Beginning to resign recovery group rg1 due to "VCD write failure", caller err 217 when "updating VCD: RGD" Note that a "VCD write failure" with err 217 is a generic message issued when fault tolerance is exceeded during critical system updates, but in this case the race condition resigns the system when only a handful of missing disks are found. (show details) |
5.1.0.2 | GNR, ESS |
IJ29815 | High Importance | "Disk in use" error when using not partitioned DASD devices. DASD '/dev/dasdk' is in use. Unmount it first! mmcrnsd: Unexpected error from fdasd -a /dev/dasd. Return code: 1 mmcrnsd: [E] Unable to partition DASD device /dev/disk/by-path/ccw-0.0.0500 mmcrnsd: Failed while processing disk stanza on node node01.abc.de %nsd: device=/dev/disk/by-path/ccw-0.0.0500 nsd=scale_data01 servers=node01.abc.de usage=dataAndMetadata (show details) | 5.1.0.2 | Installation toolkit |
IJ29917 | Suggested | When a user starts the mmrestripefile command against a big file with -b option, it could take a long time(e.g. more than 20 minutes) to return but no data movement is seen between disks. This is because the big file is already balanced. (show details) | 5.1.0.2 | mmrestripefile |
IJ29918 | High Importance | Node crash if tremendous parallel access to file with NFS (show details) | 5.1.0.2 | kNFS |
IJ29829 | Suggested | Several RAS events had inconsistent values in their SEVERITY and STATE. For instance, the event "network_bond_degraded", which STATE=DEGRADED, has SEVERITY=INFO. As a result, related failures were not propagated properly. (show details) | 5.1.0.2 | System health, GUI |
IJ29857 | High Importance | Starting in version 5.1, the afmIOFlags configuration on the fileset level is printed in Hex format and checking it in the integer range ends up printing an error to the logs. (show details) | 5.1.0.2 | AFM |
IJ29555 | Suggested | In a mirrored disk environment, I/O is expected to continue with the surviving disk if a disk experiences a problem. In case of a disk being created with a recovery group vdisk, and the recovery group is in a state that it continues to resign due to some vdisk's fault tolerance exceeded after a successful recovery, a race condition exists which could cause the logic of checking this state to be skipped. As a result of this, I/O will continue to be retried to the problem disk instead of moving on to the surviving disks. (show details) | 5.1.0.2 | GNR |
IJ29749 | Medium Importance | While migrating a file to the cloud, the GPFS daemon might hit a signal in StripeGroup::decNumAccessRights() (show details) | 5.1.0.2 | TCT |
IJ29883 | High Importance | The GPFS daemon could fail with logAssertFailed: fromNode != regP->owner. This could occur when a file system's disk configuration is changed just as a new file system manager is taking over. (show details) | 5.1.0.2 | Core GPFS |
IJ29938 | Suggested | skipRecall config does not work. (show details) | 5.1.0.2 | DMAPI |
IJ29884 | Suggested | AIO operations on encrypted files are handled as buffered IO, further decreasing the performance of the AIO operation in addition to the crytographic overhead introduced by the encryption of files in the file system. (show details) | 5.1.0.2 | Encryption |
IJ29909 | Suggested | The user action of some health events for unconfigured performance monitoring sensors contain a wrong command. (show details) | 5.1.0.2 | System Health, AFM, QoS, NFS, SMB |
IJ29942 | Suggested | In a kernel > 4.10 and file-sizes being a multiple of page sizes, a false error is returned once the read offset reaches file size. (show details) | 5.1.0.2 | Core GPFS |
IJ29943 | Suggested | When there are many threads doing sync writes through the same file descriptor, a contention on inodeCacheOjbMutex between them could impact the performance of writes. (show details) | 5.1.0.2 | Sync writes |
IJ29960 | Critical | AFM gateway nodes run out of memory during resync. glibc is known to use as many arenas as 8 times the number of CPU threads a systems has. This makes a multi-threaded program like AFM which allocates memory for queues to use a lot more memory than actually needed. (show details) | 5.1.0.2 | AFM |
IJ29989 | Suggested | The tsfindinode utility incorrectly reports file path as not found for valid inodes. (show details) | 5.1.0.2 | tsfindinode |
IJ30138 | High Importance | logAssertFailed: ofP->isInodeValid() at mnUpdateInode when doing stat() or gpfs_statlite() (show details) | 5.1.0.2 | Core GPFS |
IJ30139 | High Importance | mmbackup could backup files unnecessary after failure. (show details) | 5.1.0.2 | mmbackup |
IJ30141 | Suggested | Inodes not getting freed after user deleted them (show details) | 5.1.0.2 | Core GPFS |
IJ30142 | Suggested | mmvdisk --replace command results in message: Location XXX contains multiple disk devices. (show details) | 5.1.0.2 | GNR, ESS |
IJ30163 | High Importance | Memory leak on file system manager node during quota revoke storm (show details) | 5.1.0.2 | Quotas |
IJ30166 | Suggested | mmsmb exportacl list doesn't show "@" of the SMB share name. (show details) | 5.1.0.2 | SMB |
IJ30131 | Suggested | Some processes may not be woken up as they should during a cluster manager change. That might lead to potential deadlocks. (show details) | 5.1.0.2 | Core GPFS |
IJ30180 | Suggested | Kernel v4.7 changes the inode ACLs cache mechanism, and GPFS (5.0.5.2+, 4.2.3.23+) does not adapt to the new kernel behaviors. The following two typical issues are observed: 1. normal user can access one file, and root removes the file access privilege from the user by chmod command => the user can still access the file 2. normal user cannot access one file, and root grants the file access privilege for the user by chmod command => the user cannot access the file either. (show details) | 5.1.0.2 | Core GPFS |
IJ30191 | High Importance | When file audit logging is enabled, events that are generated on non-file system manager nodes will fail to be logged to the audit log. (show details) | 5.1.0.2 | File audit logging |
IJ30224 | Suggested | The mmfs.log shows several "sdrServ: Communication error" messages. (show details) | 5.1.0.2 | System health |
IJ30248 | High Importance | mmfsd crashed due to signal 11 when verifying the file system descriptor. (show details) | 5.1.0.2 | Core GPFS |
IJ30337 | Suggested | This problem involves adding, deleting, or changing quorum nodes when GPFS is not running on the majority of quorum nodes. Commands may be left behind a mmRunningCommand lock which prevents GPFS from starting up. (show details) | 5.1.0.2 | Admin commands |
IJ30337 | High Importance | The IBM Spectrum Scale HDFS Transparency connector version 3.1.0-6 contains 2 NullPointerExceptions in the HDFS NameNode service. The application accessing the data is not impacted, but these exceptions are seen in the NameNode log file. (show details) | 5.1.0.2 | HDFS Connector |
IJ30336 | Suggested | The IBM Spectrum Scale HDFS Transparency connector version 3.1.0-6 modified the label for the open operation when the configuration is set to "Scale" for the ranger.enabled parameter. When retrieving the JMX stats, the open is reported as GetBlockLocations. (show details) | 5.1.0.2 | HDFS Connector |
IJ29025 | High Importance | Given an IBM Spectrum Scale cluster with 'verbsRdmaCm' set to 'enable' and configured to use RDMA through RoCE, individual nodes will fail to establish an RDMA connection to other nodes when the IP addresses configured on the RDMA adapters include a non-link local IPv6 address. (show details) | 5.1.0.1 | RDMA |
IJ29026 | High Importance | Under heavy workload (especially file creation/deletion involved) with quota function enabled, some race issues are exposed such that the filesetId is not handled correctly, causing a GPFS daemon assert. (show details) | 5.1.0.1 | Quotas |
IJ29027 | Suggested | If the cluster is configured with separate daemon and admin interfaces, the -Y output of mmgetstate only shows the admin node name. (show details) | 5.1.0.1 | Admin commands |
IJ29154 | High Importance | mmfsd daemon assert going off: Assert exp(rmsgP != __null) in file llcomm.C, resulting in a daemon crash. (show details) | 5.1.0.1 | Core GPFS |
IJ29155 | Suggested | While a node is tryiing to join a cluster, mmfsd start could encounter a null pointer reference and crash with a signal 11 with a backstack that looks like this: [D] #0: 0x0000559601506BCE RGMaster::getNode FullDomainName(NodeAddr, char**) + 0xAE at ??:0 [D] #1: 0x000055960150CAA2 RGMaster::rgListServers(int, unsigned int) + 0x212 at ??:0 [D] #2: 0x000055960145F21C runTSLsRecoveryGroupV2 (int, StripeGroup*, int, char**) + 0xA8C at ??:0 [D] #3: 0x0000559601460371 runTSLsRecoveryGroup (int, StripeGroup*, int, char**) + 0xB1 at ??:0 (show details) |
5.1.0.1 | GNR |
IJ29157 | High Importance | logAssertFailed: !"Cleanup hit contended Fileset lock." (show details) | 5.1.0.1 | Filesets |
IJ29028 | Suggested | mmlsmount with the --report and -Y options may not take into account nodes which do not have the file system mounted. (show details) | 5.1.0.1 | Core GPFS |
IJ29045 | Suggested | mmhealth cluster show: faster heartbeat_missing (show details) | 5.1.0.1 | System health |
IJ29161 | Suggested | The mmkeyserv command displays the latest expiration date from the KMIPT certificate chain. It should display the expiration date of the end-entity certificate. (show details) | 5.1.0.1 | Core GPFS, Admin commands, Encryption |
IJ29162 | Suggested | The systemhealth monitor reports data and name nodes as down for the HadoopConService. In fact, both were running. (show details) | 5.1.0.1 | System health |
IJ29163 | Critical | When truncating a migrated immutable file with DMAPI interfaces, the data of the file becomes zero, although the file is immutable. (show details) | 5.1.0.1 | Immutable and append-only files |
IJ29064 | High Importance | When file audit logging or watch folder is enabled on a file system, unmounting the file system might result in a waiter that will not clear. This may cause other commands to hang. (show details) | 5.1.0.1 | File audit logging, Watch folder |
IJ29186 | High Importance | On clusters having minReleaseLevel at 5.0.1, with mixed version nodes available from 5.0.1.x through till 5.0.5.x nodes and where the gateway node is at level 5.0.5.x - the newer level of gateway nodes finds it hard to co-exist with the older level nodes causing a recovery failure repeatedly. (show details) | 5.1.0.1 | AFM |
IJ28848 | Suggested | Cannot create SMB share using utf8 chars through CLI. (show details) | 5.1.0.1 | SMB |
IJ29134 | Medium Importance | The systemhealth monitor reported a gpfs_down event and triggered a failover even though the system was fine. (show details) | 5.1.0.1 | System health |
IJ29135 | Critical | mmhealth does not work on AIX. (show details) | 5.1.0.1 | System health |
IJ29002 | Critical | If the default replication (-m or -r) setting for a file system is set to 1, and mmvdisk is used to add an additional vdisk set to the file system, an exception will be hit if the --failure-groups option is not used. (show details) | 5.1.0.1 | ESS, GNR |
IJ29210 | High Importance | mmvdisk recovery group fails when creating log vdisks when creating a new recovery group in a cluster with preexisting recovery groups. An error message "Disk XXX is already registered for use by GPFS" will appear on the command console, and the recovery group creation will fail. Once the problem condition is hit, IBM support must be contacted to correct the conflicting cluster information. (show details) | 5.1.0.1 | Admin commands, ESS, GNR |
IJ29213 | Suggested | While reading the file, the file can be evicted and its captured checksum shows inconsistency for this opened file. (show details) | 5.1.0.1 | AFM |
IJ29214 | High Importance | If a hostname is resolved to a loopback address, the CCR component might run into an assertion of type 'ccrNodeId == iter->id' when a node becomes quorum node using the mmchnode or mmaddnode command. (show details) | 5.1.0.1 | Admin commands, CCR |
IJ29216 | Critical | After mmimgrestore, the mmfsd could assert when handling the mmlsfileset command for a dependent fileset: logAssertFailed: fsOfP->getDirLayoutP() != __null (show details) | 5.1.0.1 | DMAPI, HSM, TSM |
IJ29248 | Suggested | IBM Spectrum Scale has core dump triggered in dAssignSharedBufferSpace() due to segmentation fault hit by mmfsd or lxtrace daemon. (show details) | 5.1.0.1 | Trace |
IJ29188 | Suggested | When the file system is in panic on a quota client node, the outstanding quota share is not relinquished. Quota share Indoubt value is reported and the shares can only be reclaimed by mmcheckquota. (show details) | 5.1.0.1 | Quotas |
IJ29190 | Suggested | Incorrect quota check results on small files with fragments (show details) | 5.1.0.1 | Quotas |
IJ29201 | Suggested | Incorrect quota check result due to OpenFile reuse/updateShadowTab (show details) | 5.1.0.1 | Quotas |
IJ29217 | Critical | --metadata-only option hit the assert Assert exp(!"Assert on Structure Error") in prefetch. (show details) | 5.1.0.1 | AFM, AFM DR |
IJ29275 | Suggested | Running prefetch stats is failing with err 22. (show details) | 5.1.0.1 | AFM |
IJ29243 | Critical | File system manager could assert with exp(isStoragePoolIdValid(poolId)) during log recovery if a node fails shortly after running mmdeldisk. (show details) | 5.1.0.1 | Core GPFS |
IJ29251 | Suggested | On a platform with NUMA support, GPFS may report no platform NUMA support is available. (show details) | 5.1.0.1 | Linux NUMA subsystem detection of system resources |
IJ29252 | High Importance | Installation, update or configuration of the object protocol on 5.1.0.0 fails with a message saying the configuration is not supported or that required dependencies cannot be found. (show details) | 5.1.0.1 | Object |
IJ29253 | High Importance | In cases which has small pagepool size and large file system block size, GPFS may wait for page reservation unnecessarily because GPFS tends to reserve more pages than necessary. (show details) | 5.1.0.1 | mmap |
IJ29255 | High Importance | While the GPFS daemon is shutting down, there is chance a specific trace will be logged and it may crash the kernel. (show details) | 5.1.0.1 | Core GPFS |
IJ29261 | High Importance | On zLinux, while running a mmap workload with traceIoData configuration enabled, the trace code may trigger a page fault and cause the kernel to crash. (show details) | 5.1.0.1 | mmap |
IJ29263 | High Importance | mmchnode fails when more than the current number of quorum nodes becomes quorum nodes again. (show details) | 5.1.0.1 | mmchnode --quorum, CCR |
IJ29356 | Critical | GPFS maintains EA (Extended Attribute) registry to verify the EA priority. Due to incorrect EA registry addition without SG format version check, policy and inode scans might fail in the mixed node cluster environment. This problem could occur while running policy or inode scans in a mixed node environment running with 5.0.5.2, 5.0.5.3, and 5.1.0.0 and other old version nodes as the file system manager. (show details) | 5.1.0.1 | AFM, Core GPFS |
IJ29377 | High Importance | When file audit logging is enabled and audit events are being generated, and a file system is panicked, the IBM Spectrum Scale node where the panic happened may assert. (show details) | 5.1.0.1 | File audit logging |
IJ29428 | Critical | When an uncached file is renamed in the local-updates mode, the file is not copied to the previous snapshot causing the setInodeDirtyAndVerify assert. (show details) | 5.1.0.1 | AFM |
IJ29417 | High Importance | If a CCR backup archive is used to restore the CCR component with the command 'mmsdrrestore -F |
5.1.0.1 | Admin commands, CCR |
IJ29676 | HIPER | The GPFS daemon on a file system manager node could fail with logAssertFailed: nFailedNodes <= 64. This could happen on a large cluster where more than 64 nodes fail around the same time. (show details) | 5.1.0.1 | Core GPFS |
IJ30425 | High Importance | mmvdisk throws an exception for a list operation when the daemon node name is not identical to the admin node name. (show details) | 5.1.0.1 | GNR, ESS |
IJ23984 | High Importance | If the directory passed to the --directory option has spaces or any special characters in the name, then prefetch is not able to handle them correctly. And, it fails printing the usage error to exit. (show details) | 5.1.0.0 | AFM |
IJ25712 | High Importance | Due to a delayed file close in the VFS layer and the context mismatch, closing the file after the replication does not wait for the file system quiesce causing the remote log assert. (show details) | 5.1.0.0 | AFM, AFM DR |
IJ25754 | High Importance | Quota clients request quota shares based on the workload and most of the time the quota shares given to an active client is much larger than the previously pre-defined amount (e.g. 20 file system blocks). The unused or excess quota shares are returned to the quota manager periodically. At the quota manager side, when the quota usage exceeds the established soft quota limits, the grace period is triggered. When this event occurs, the quota shares are reclaimed and the quota share distribution falls back to a more conservative fashion (based on a predetermined amount). In certain workloads, when the partial quota shares are returned to the manager along with the usage updates and as a result it triggers the soft quota limit exceeded event, some amount of quota shares are lost due to mismanagement of quota shares between the client and the manager. This leads to permanent loss of quota shares correctable by the mmcheckquota command. (show details) | 5.1.0.0 | Quotas |
IJ25802 | High Importance | There is no automatic method to generate the GPL installable package for customized Red Hat Enterprise Linux releases. (show details) | 5.1.0.0 | Core GPFS |
IJ25532 | High Importance | The snapshot deletion command fails with error 784(E_ENC_KEY_UNMARSHAL). This is because one of the snapshots file encryption attribute is corrupted. (show details) | 5.1.0.0 | Snapshots, Encryption |
IJ26654 | High Importance | The message csm_resync_needed should only show up if the communication between a node and the cluster manager was broken for a given amount of time. In mixed version environments, the events "csm_resync_needed" and "heartbeat_missing" messages may be shown erroneously. (show details) | 5.1.0.0 | System health |
IJ26520 | Suggested | mmhealth event hide might take a long time to be included in the state as some checks are done very infrequently. (show details) | 5.1.0.0 | System health |
IJ27923 | Suggested | When a user turns off the file system maintenance mode, the file system cannot be mounted. (show details) | 5.1.0.0 | Core GPFS |
IJ28607 | High Importance | afmFastCreate when set doesn't filter create messages already in the queue when remove comes in. As a result, both a create and a remove on same file can exist on the queue at the same time. If a link gets sandwiched between the create and the remove then the link fails to find the remote file and sees it as a conflict and drops the queue. Causing resync + recovery later to complete the sync. (show details) | 5.1.0.0 | AFM, AFM DR |
IJ27087 | High Importance | Application runs with I/O priority mapping into a not supported QoS class, which does have IOPS limitation with 1 IOPS. Thus, leading to I/Os being queued to wait for enough tokens to service the I/O operation. This causes long waiters. (show details) | 5.1.0.0 | QoS |
IJ28608 | High Importance | If call home data collection process was interrupted because of a the power loss, the following data collection of the same schedule will fail due to the directory already existing. (show details) | 5.1.0.0 | Call home |