HU02327 |
All |
HIPER
|
Using addvdiskcopy in conjunction with expandvdisk with format may result in the original being overwritten, by the new copy, producing blank copies. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems running v8.2.1 or later |
Trigger |
Using addvdiskcopy and expandvdisk with format |
Workaround |
Wait until the format is completed before adding a copy |
|
8.2.1.15 |
Volume Mirroring |
HU02400 |
All |
HIPER
|
A problem in the virtualization component of the system can cause a migration IO to be submitted in an incorrect context resulting in a node warmstart. In some cases it is possible that this IO has been submitted to an incorrect location on the backend, which can cause data corruption of an isolated small area
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems running v8.2.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.15 |
Storage Virtualisation |
HU02342 |
All |
Critical
|
Occasionally when an offline drive returns to online state later than its peers in the same RAID array there can be multiple node warmstarts that send nodes into a service state
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.15 |
RAID |
HU02374 |
SVC, V5000, V7000 |
Critical
|
Hosts with Emulex 16Gbps HBAs may become unable to communicate with a system with 8Gbps Fibre Channel ports, after the host HBA is upgraded to firmware version 12.8.364.11. This does not apply to systems with 16Gb or 32Gb Fibre Channel ports
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with 8Gbps Fibre Channel ports |
Trigger |
Host Emulex 16Gbps HBA is upgraded to firmware version 12.8.364.11 |
Workaround |
Do not upgrade host HBA to firmware version 12.8.364.11 |
|
8.2.1.15 |
Hosts |
HU02393 |
All |
Critical
|
Automatic resize of compressed/thin volumes may fail causing warmstarts on both nodes in an I/O group
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.15 |
Storage Virtualisation |
HU02401 |
All |
Critical
|
EasyTier can move extents between identical mdisks until one runs out of space
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1 or later using EasyTier |
Trigger |
None |
Workaround |
Disable EasyTier. Manually migrate extents between mdisks |
|
8.2.1.15 |
EasyTier |
HU02406 |
All |
Critical
|
An interoperability issue between Cisco NX-OS firmware and the Spectrum Virtualize Fibre Channel driver can cause a node warmstart on NPIV failback (for example during an upgrade) with the potential for a loss of access. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using NPIV that are connected to Cisco SAN equipment running NX-OS 8.4(2c) or later |
Trigger |
Initiate an NPIV failback operation by, for example, performing an upgrade |
Workaround |
Disable NPIV (which will require any hot spare nodes to be removed first) |
|
8.2.1.15 |
Interoperability |
IT38015 |
All |
High Importance
|
During RAID rebuild or copyback on systems with 16gb or less of memory, cache handling can lead to a deadlock which results in timeouts
(show details)
Symptom |
Performance |
Environment |
Systems with 16GB or less of memory |
Trigger |
None |
Workaround |
None |
|
8.2.1.15 |
RAID |
HU02366 |
All |
Suggested
|
Slow internal resource reclamation by the RAID component can cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.2.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.15 |
RAID |
HU02433 |
FS5000, SVC, V5000, V5100, V7000 |
Suggested
|
When a BIOS upgrade occurs excessive tracefile entries can be generated
(show details)
Symptom |
None |
Environment |
Gen 1 & 2 systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.15 |
System Update |
HU02186 |
FS9100, V5100, V7000 |
HIPER
|
NVMe drive pulls or firmware upgrades may lead to offline pools with the possibility of a small loss of data integrity. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems running v8.2.1, or later with NVMe drives |
Trigger |
None |
Workaround |
None |
|
8.2.1.13 |
RAID |
HU02186 (reverted) |
FS9100, V5100, V7000 |
HIPER
|
This APAR has been reverted at this PTF. This APAR will be re-applied in a future PTF
|
8.2.1.12 |
RAID |
HU02277 |
All |
HIPER
|
RAID parity scrubbing can become stalled causing an accumulation of media errors leading to multiple drive failures with the possibility of data integrity loss. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems with model MZILS3T8HMLH read intensive SSDs at drive firmware MS24 are particularly susceptible to the data integrity (DI) issue. Other drive types may see multiple failures without DI issue |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
RAID |
HU02313 |
FS9100, V5100, V7000 |
HIPER
|
When a FlashCore Module (FCM) fails there is a chance that this can trigger other FCMs in the same control enclosure to also fail. If enough additional drives fail, at the same time, this can take the array offline and cause a loss of access to data. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2 or later using Flash Core Modules |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Drives |
HU01968 & HU02215 |
All |
Critical
|
An upgrade may fail due to corrupt hardened data in a node. This can affect an I/O group
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
System Update |
HU02058 |
All |
Critical
|
Changing a remote copy relationship from GMCV to MM or GM can result in a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using remote copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02184 |
All |
Critical
|
When a 3PAR controller experiences a fault that prevents normal I/O processing it may issue a SCSI TARGET RESET command. This command is not supported and may cause multiple node asserts, possibly cluster-wide
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with 3PAR backend controllers |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Backend Storage |
HU02213 |
SVC |
Critical
|
A Hot Spare Node (HSN) timing window issue can, during an HSN activation or deactivation, cause the cluster to broadcast an invalid VPD update to other clusters on the SAN. This may trigger a Tier 2 recovery on the other cluster. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
SVC systems, with Hot Spare Nodes, using remote copy partnerships |
Trigger |
None |
Workaround |
Prior to an upgrade, or node hardware maintenance, remove the HSN |
|
8.2.1.12 |
Hot Spare Node |
HU02266 |
All |
Critical
|
An issue in auto-expand can cause expansion to fail and the volume to be taken offline
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v8.2.1 or later using thin-provisioning |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Thin Provisioning |
HU02295 |
SVC |
Critical
|
When upgrading from v8.2.1 or v8.3, in the presence of hot spare nodes, an issue with the handling of node metadata may cause a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
SVC systems running v8.2.1 or v8.3 with Hot Spare Node |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
System Update |
HU02429 |
All |
Critical
|
System can go offline shortly after changing the SMTP settings using the chemailserver command via the GUI
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
System Monitoring |
HU02132 |
All |
High Importance
|
Removing a thin-provisioned volume and then immediately creating one of the same size may cause node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using thin-provisioned volumes |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Thin Provisioning |
HU02156 |
All |
High Importance
|
Global Mirror environments may experience more frequent 1920 events due to writedone message queuing
(show details)
Symptom |
Performance |
Environment |
Systems using Global Mirror |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Global Mirror |
HU02164 |
All |
High Importance
|
An issue in Remote Copy may cause a loss of hardened data when a node is warmstarted
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems using remote copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02176 |
All |
High Importance
|
During upgrade a node may limit the number of target ports it reports causing a failover contradiction on hosts
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with 12 or more Fibre Channel ports presenting storage to host type "tpgs" |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Hosts |
HU02200 |
All |
High Importance
|
When upgrading from v8.1 or earlier to v8.2.1 or later a remote copy issue may cause a node warmstart, stalling the upgrade
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.1 or earlier using remote copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
System Update |
HU02201 & HU02221 |
All |
High Importance
|
Shortly after upgrading drive firmware, specific drive models can fail due to Too many long IOs to drive for too long errors
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with the following drive models:
- ST300MM0009 (300GB) - B5B8
- ST600MM0009 (600GB) - B5B8
- ST900MM0009 (900GB) - B5B8
- ST1200MM0009 (1200GB) - B5B8
- ST1200MM0129 (1800GB) - B5C9
- ST2400MM0129 (2400GB) - B5C9
- ST300MP0006 (300GB) - B6AA
- ST600MP0006 (600GB) - B6AA
- ST900MP0146 (900GB) - B6CB
|
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Drives |
HU02227 |
FS9100, SVC, V5100, V7000 |
High Importance
|
Certain I/O patterns can cause compression hardware to post errors. When those errors exceed a threshold the node can be taken offline
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v8.2 or later using compressed volumes |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Compression |
HU02288 |
All |
High Importance
|
A node might fail to come online after a reboot or warmstart such as during an upgrade
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems in a Stretched or HyperSwap topology |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Reliability Availability Serviceability |
HU02048 |
All |
Suggested
|
An issue in the handling of ATS commands from VMware hosts can cause a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8, or later, presenting volumes to VMware hosts |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Hosts |
HU02095 |
All |
Suggested
|
The effective_used_capacity field of lsarray/lsmdisk commands should be empty for RAID arrays which do not contain overprovisioned drives. However, sometimes this field can be zero even though it should be empty. This can cause incorrect provisioned capacity reporting in the GUI
(show details)
Symptom |
None |
Environment |
Systems with non-FCM arrays |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Graphical User Interface |
HU02142 |
All |
Suggested
|
It is possible for a backend unmap process to become stalled, preventing system configuration changes from completing
(show details)
Symptom |
Configuration |
Environment |
Systems running v8.1.0, or later, using DRAID |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Distributed RAID |
HU02157 |
All |
Suggested
|
Issuing a mkdistributedarray command may result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Distributed RAID |
HU02241 |
All |
Suggested
|
IP Replication can fail to create IP partnerships via the secondary cluster management IP
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
Use primary management IP to run mkippartnership commands |
|
8.2.1.12 |
IP Replication |
HU02244 |
SVC |
Suggested
|
False positive node error 766 (depleted CMOS battery) messages may appear in the Event Log
(show details)
Symptom |
None |
Environment |
SVC systems with SV1 model nodes running v8.2.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
System Monitoring |
HU02292 & HU02308 |
All |
Suggested
|
The use of maximum replication delay within Global Mirror may occasionally cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Remote Copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Global Mirror |
HU02332 & HU02336 |
All |
Suggested
|
When an I/O is received, from a host, with invalid or inconsistent SCSI data but a good checksum it may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Hosts |
HU02354 |
SVC |
Suggested
|
An issue in the handling of read transfers may cause hung host IOs leading to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.2.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Hosts |
HU02358 |
All |
Suggested
|
An issue in Remote Copy, that stalls a switch of direction, can cause I/O timeouts leading to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Remote Copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.12 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU01894 |
All |
HIPER
|
After node reboot, or warmstart, some volumes accessed by AIX, VIO or VMware hosts may experience stuck SCSI2 reservations on the NPIV failover ports of the partner node. This can cause a loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using NPIV to present storage to AIX, VIO or VMware hosts |
Trigger |
None |
Workaround |
Clear reservation by either:
- Unmap & re-map volume;
- LUN reset from host.
|
|
8.2.1.11 |
Hosts |
HU02141 |
All |
HIPER
|
An issue in the max replication delay function may trigger a Tier 2 recovery, after posting multiple 1920 errors in the Event Log. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using remote copy |
Trigger |
None |
Workaround |
Set the max_replication_delay value to 0 (disabled) |
|
8.2.1.11 |
Global Mirror |
HU02186 (reverted in 8.2.1.12) |
FS9100, V5100, V7000 |
HIPER
|
NVMe drive pulls or firmware upgrades may lead to offline pools with the possibility of a small loss of data integrity. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems running v8.2.1, or later with NVMe drives |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
RAID |
HU02205 |
All |
HIPER
|
Incremental FlashCopy targets can be corrupted when the FlashCopy source is a target of a remote copy relationship
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems using Incremental FlashCopy with remote copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
FlashCopy, Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02212 |
All |
HIPER
|
Remote Copy secondary may have inconsistent data following a stop with -access due to a missing bitmap merge from FlashCopy to Remote Copy. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems running v8.2.1 or later using GMCV or HyperSwap |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Global Mirror With Change Volumes, HyperSwap |
HU02237 |
All |
HIPER
|
Under a rare and complicated set of conditions, a RAID 1 or RAID 10 array may drop a write, causing undetected data corruption. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems using RAID 1 or RAID 10 arrays |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
RAID |
HU02238 |
All |
HIPER
|
Force-stopping a FlashCopy map, where the source volume is a Metro or Global Mirror target volume, may cause other FlashCopy maps to return invalid data if they are not 100% copied, in specific configurations. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
FlashCopy, Global Mirror, Metro Mirror |
HU01921 |
All |
Critical
|
Where FlashCopy mapping targets are also in remote copy relationships there may be node warmstarts with a temporary loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using FlashCopy with remote copy |
Trigger |
None |
Workaround |
If one reverse FlashCopy mapping has been stopped and another FlashCopy mapping, to the same target, is to be started, then delete the first reverse FlashCopy mapping before starting the second |
|
8.2.1.11 |
FlashCopy, Global Mirror, Metro Mirror |
HU01924 |
All |
Critical
|
Migrating extents to an MDisk, that is not a member of an MDisk group, may result in a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Migrate extents to an MDisk, that is not a member of an MDisk group |
Workaround |
Only specify a target MDisk that is part of the same MDisk group as the volume copy having extents migrated |
|
8.2.1.11 |
Thin Provisioning |
HU01970 |
All |
Critical
|
When a GMCV relationship is stopped, with the -access option, and the secondary volume is immediately deleted with -force, then all nodes may repeatedly warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using GMCV |
Trigger |
Stop a GMCV relationship with -access and immediately delete the secondary volume |
Workaround |
Do not remove secondary volume, with -force, if the backward FC map from the secondary change volume to the secondary volume is still in progress |
|
8.2.1.11 |
Global Mirror With Change Volumes |
HU02005 |
All |
Critical
|
An issue in the background copy process prevents grains, above a 128TB limit, from being cleaned properly. As a consequence there may be multiple node warmstarts with the potential for a loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using remote copy |
Trigger |
Volumes greater than 128TB being added to remote copy relationships |
Workaround |
The maximum size of volumes in remote copy relationships should be limited to 128TB |
|
8.2.1.11 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02054 |
All |
Critical
|
The event log handler maintains a second list of events. On rare occasions, for log full events, these lists can get out of step, resulting in a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
If an error event log full message (1002) is presented clear the event log, rather than marking that event as fixed |
|
8.2.1.11 |
System Monitoring |
HU02065 |
All |
Critical
|
Mishandling of Data Reduction Pool allocation request rejections can lead to node warmstarts that can take an MDisk group offline
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Data Reduction Pools |
HU02069 |
All |
Critical
|
When a SCSI command, containing an invalid byte, is received there may be a node warmstart. This can affect both nodes, in an I/O group, at the same time
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with hosts using Fibre Channel connectivity |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Hosts |
HU02097 |
All |
Critical
|
Workloads, with data that is highly suited to deduplication, can provoke high CPU utilisation, as multiple destinations try to dedupe to one source. This adversely impacts performance with the possibility of offline MDisk groups
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Data Reduction Pools |
HU02106 |
All |
Critical
|
Multiple node warmstarts, in quick succession, can cause the partner node to lease expire
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using IP Quorum or NVMe drives as quorum devices |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
IP Quorum, Quorum |
HU02108 |
All |
Critical
|
Deleting a managed disk group, with -force, may cause multiple warmstarts with the possibility of a loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Data Reduction Pools |
HU02109 |
All |
Critical
|
Free extents may not be unmapped after volume deletion, or migration, resulting in out-of-space conditions on backend controllers
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.3.0 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Backend Storage, SCSI Unmap |
HU02135 |
All |
Critical
|
Removing multiple IQNs for an iSCSI host can result in a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2 or later with iSCSI connected hosts |
Trigger |
Use a single rmhostport command to remove multiple IQN from an iSCSI host |
Workaround |
Remove iSCSI host ports one IQH at a time |
|
8.2.1.11 |
iSCSI |
HU02138 |
All |
Critical
|
An issue in Data Reduction Pool garbage collection can cause I/O timeouts leading to an offline pool
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Data Reduction Pools |
HU02154 |
All |
Critical
|
If a node is rebooted, when remote support is enabled, then all other nodes will warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1 or later using remote support |
Trigger |
With remote support enabled, reboot a node using the 'satask stopnode -reboot <node id>' command |
Workaround |
Temporarily disable remote support when rebooting a node using 'chsra -remotesupport disable' |
|
8.2.1.11 |
Support Remote Assist |
HU02155 |
All |
Critical
|
Upgrading to v8.2.1 may result in offline managed disk groups and OOS events (1685/1687) appearing in the Event Log
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2.0, or earlier, using Data Reduction Pools |
Trigger |
Upgrading to v8.2.1 |
Workaround |
None |
|
8.2.1.11 |
Data Reduction Pools |
HU02197 |
All |
Critical
|
Bulk volume removals can adversely impact related FlashCopy mappings leading to a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1, or later, using FlashCopy |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
FlashCopy |
HU02222 |
All |
Critical
|
Where the source volume of an incremental FlashCopy map is also a Metro or Global Mirror target volume that is using a change volume or is a Hyperswap volume, then there is a possibility that not all data will be copied to the FlashCopy target. For more details refer to this Flash
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems using Remote Copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Global Mirror With Change Volumes |
IT25367 |
All |
Critical
|
A T2 recovery may occur when an attempt is made to upgrade, or downgrade, the firmware for an unsupported drive type
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Attempt to upgrade/downgrade the firmware for an unsupported drive type |
Workaround |
None |
|
8.2.1.11 |
Drives |
IT31113 |
All |
Critical
|
After a manual power off and on, of a system, both nodes, in an I/O group, may repeatedly assert into a service state
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2 or later |
Trigger |
Manual power off and on of a system whilst a RAID rebuild is in progress |
Workaround |
None |
|
8.2.1.11 |
RAID |
HU01923 |
All |
High Importance
|
An issue in the way Global Mirror handles write sequence numbers >512 may cause multiple node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using GM |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Global Mirror |
HU02080 |
All |
High Importance
|
When a Data Reduction Pool is running low on free space, the credit allocation algorithm, for garbage collection, can be exposed to a race condition, adversely affecting performance
(show details)
Symptom |
Performance |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Data Reduction Pools |
HU02114 |
FS5000, FS9100, V7000 |
High Importance
|
Upgrading FCM firmware on multiple I/O group systems can cause a drive to become stuck at 0% sync with the corresponding array in a 'syncing' state
(show details)
Symptom |
Performance |
Environment |
Multiple I/O group systems with Flash Core Modules |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Drives |
HU02123 |
All |
High Importance
|
For direct-attached hosts, a race condition between the FLOGI and Link UP processes can result in FC ports not coming online
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with direct-attached hosts |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Hosts |
HU02149 |
SVC |
High Importance
|
When an Enhanced Stretch Cluster is using NPIV, in transitional mode, the path priority is not being reported correctly to some hosts
(show details)
Symptom |
Performance |
Environment |
Systems in an Enhanced Stretch Cluster topology that are using NPIV in its transitional mode |
Trigger |
None |
Workaround |
Manually set the preferred path if possible within the host's MPIO settings. Run NPIV in enabled or disabled mode |
|
8.2.1.11 |
Hosts |
HU02168 |
V5000, V7000 |
High Importance
|
In the event of unexpected power loss a node may not save system data
(show details)
Symptom |
Loss of Redundancy |
Environment |
Storwize V5000 Gen2, V7000 Gen 2 and Gen 2+ systems |
Trigger |
Sudden power loss |
Workaround |
When shutting down always use the CLI, service GUI or management GUI. Do not use removal of electrical supply |
|
8.2.1.11 |
Reliability Availability Serviceability |
HU02203 |
FS9100, V5000, V7000 |
High Importance
|
When a node reboots, it is possible for the node to be unable to communicate with some of the NVMe drives in the enclosure
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with NVMe drives |
Trigger |
Drive reseat or node reboot |
Workaround |
None |
|
8.2.1.11 |
Drives |
HU01868 |
All |
Suggested
|
After deleting an encrypted external MDisk, it is possible for the encrypted status of volumes to change to no, even though all remaining MDisks are encrypted
(show details)
Symptom |
None |
Environment |
Systems using encryption |
Trigger |
Delete an encrypted external MDisk |
Workaround |
Ensure that all MDisks in the MDisk group are encrypted - this will ensure that data is encrypted |
|
8.2.1.11 |
Encryption |
HU01917 |
All |
Suggested
|
Chrome browser support requires a self-signed certificate to include subject alternate name
(show details)
Symptom |
None |
Environment |
Systems accessed using the Chrome browser |
Trigger |
None |
Workaround |
Accept invalid certificate |
|
8.2.1.11 |
Graphical User Interface |
HU01931 |
SVC, V7000 |
Suggested
|
Where a high rate of CLI commands are received, it is possible for inter-node processing code to be delayed which results in a small increase in receive queue time on the config node
(show details)
Symptom |
Performance |
Environment |
SVC and Storwize V7000 systems |
Trigger |
None |
Workaround |
If CPU utilisation is less than 40% then creating a compressed volume may reduce response times |
|
8.2.1.11 |
Performance |
HU02015 |
FS9100, V5000, V7000 |
Suggested
|
Some read-intensive SSDs are incorrectly reporting wear rate thresholds generating unnecessary errors in the Event Log
(show details)
Symptom |
None |
Environment |
Systems using Toshiba M4 Read-Intensive SSDs |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Drives |
HU02091 |
V5000 |
Suggested
|
Upgrading to v8.2.1.8, or later, may result in a licensing error in the Event Log
(show details)
Symptom |
None |
Environment |
Lenovo Storage V Series systems |
Trigger |
Upgrade to v8.2.1.8 or later |
Workaround |
None |
|
8.2.1.11 |
Licensing |
HU02103 |
FS9100, V5000, V7000 |
Suggested
|
The system management firmware may, incorrectly, attempt to obtain an IP address, using DHCP, making it accessible via Ethernet
(show details)
Symptom |
None |
Environment |
FlashSystem 9100, Storwize V7000 Gen 3 and Storwize V5100 systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
|
HU02111 |
All |
Suggested
|
An issue with how Data Reduction Pool handles data, at the sub-extent level, may result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Data Reduction Pools |
HU02124 |
All |
Suggested
|
Due to an issue with FCM thin provisioning calculations the GUI may incorrectly display volume capacity and capacity savings as zero
(show details)
Symptom |
None |
Environment |
Systems running v8.2.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
System Monitoring |
HU02137 |
All |
Suggested
|
An issue with support for target resets in Nimble Storage controllers may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.2, or later, with Nimble Storage backend controllers |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Backend Storage |
HU02173 |
All |
Suggested
|
During a pending fabric login, when an abort is received, it is possible for a related entry in the WWPN table to not be removed. The node will warmstart to clear this condition
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.2 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Reliability Availability Serviceability |
HU02183 |
All |
Suggested
|
An issue in the way inter-node communication is handled can lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.2 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Reliability Availability Serviceability |
HU02190 |
All |
Suggested
|
Error 1046 not triggering a Call Home even though it is a hardware fault
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
System Monitoring |
HU02214 |
All |
Suggested
|
Under a certain I/O pattern it is possible for metadata management in Data Reduction Pools to become inconsistent leading to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Data Reduction Pools |
HU02247 |
All |
Suggested
|
Unnecessary Ethernet MAC flapping messages reported in switch logs
(show details)
Symptom |
None |
Environment |
FlashSystem 9100, Storwize V7000 Gen 3 and V5100 systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.11 |
Reliability Availability Serviceability |
HU02143 |
All |
High Importance
|
The performance profile, for some enterprise tier drives, may not correctly match the drives capabilities leading to that tier being overdriven
(show details)
Symptom |
Performance |
Environment |
Systems running v8.2 or later using EasyTier. Note: This issue does not affect DRAID 5 arrays with stripe width of 8 or 9, or DRAID6 arrays with stripe width of 10 or 12. |
Trigger |
None |
Workaround |
None |
|
8.2.1.10 |
EasyTier |
HU02104 |
All |
HIPER
|
An issue in the RAID component, in the presence of very high I/O workload and the exhaustion of cache resources, can see a deadlock condition occurring which prevents further I/O processing. The system detects this issue and takes the storage pool offline for a six minute period, to clear the problem. The pool is then brought online automatically, and normal operation resumes. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2.1 or later |
Trigger |
None |
Workaround |
Consider using a pool throttle to limit the I/O throughput |
|
8.2.1.9 |
RAID |
HU02133 |
FS9100, V5000, V7000 |
High Importance
|
NVMe drives may become degraded after a drive reseat or node reboot
(show details)
Symptom |
None |
Environment |
Systems with NVMe drives |
Trigger |
Drive reseat or node reboot |
Workaround |
None |
|
8.2.1.9 |
Drives |
HU02102 |
All |
Suggested
|
Excessive processing time required for FlashCopy bitmap operations, associated with large (> 20TB) Global Mirror change volumes, may lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Global Mirror with Change Volumes where some GMCV volumes are >20TB |
Trigger |
None |
Workaround |
Limit GMCV volume capacity to 20TB or less |
|
8.2.1.9 |
Global Mirror With Change Volumes |
HU02126 |
SVC, V5000, V7000 |
Suggested
|
There is a low probability that excessive SSH connections may trigger a single node warmstart on the configuration node
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems with Gen 1 and 2 hardware |
Trigger |
More than one SSH connection attempt per second might occasionally cause the config node to warmstart |
Workaround |
Reduce frequency of SSH connections |
|
8.2.1.9 |
Command Line Interface |
HU02131 |
All |
Suggested
|
When changing DRAID configuration, for an array with an active workload, a deadlock condition can occur resulting in a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.2.1, or later, using DRAID |
Trigger |
None |
Workaround |
None |
|
8.2.1.9 |
Distributed RAID |
HU02064 |
SVC, V7000 |
HIPER
|
An issue in the firmware for compression accelerator cards can cause offline compressed volumes. For more details refer to this Flash
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v8.2.1.x, or later, using hardware compression |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Compression |
HU02083 |
All |
HIPER
|
During DRAID rebuilds, an issue in the handling of memory buffers can lead to multiple node warmstarts and a loss of access to data. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2.1.6 or v8.3.0.0 using DRAID. Probability is highest for systems with an exact multiple of 48 drives and a stripe width of 16 |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Distributed RAID |
HU01967 |
All |
Critical
|
When I/O, in remote copy relationships, experiences delays (1720 and/or 1920 errors are logged) an I/O group may warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using remote copy |
Trigger |
Performance issues affecting replication I/O |
Workaround |
Use a max replication delay value of 30 seconds or greater |
|
8.2.1.8 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02036 |
All |
Critical
|
It is possible for commands, that alter pool-level extent reservations (i.e. migratevdisk or rmmdisk), to conflict with an ongoing EasyTier migration, resulting in a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2 or later with EasyTier enabled |
Trigger |
None |
Workaround |
Disable EasyTier on the source pool; Wait 10 minutes, so that any ongoing EasyTier requests complete; Issue the migratevdisk/rmmdisk command; Enable EasyTier on the source pool. |
|
8.2.1.8 |
EasyTier |
HU02044 |
All |
Critical
|
Multiple DRAID arrays can, where one is performing a rebuild, be exposed to a RAID deadlock condition resulting in multiple node warmstarts and a loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Distributed RAID with Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Data Reduction Pools, Distributed RAID |
HU02050 |
FS9100, V5000, V7000 |
Critical
|
Compression hardware can have an issue processing certain types of data resulting in node reboots and marking the compression hardware as faulty even though it is serviceable
(show details)
Symptom |
Loss of Access to Data |
Environment |
FlashSystem 9100, Storwize V5100 and V7000 Gen 3 systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Compression |
HU02063 |
All |
Critical
|
HyperSwap clusters with only two surviving nodes may experience warmstarts on both of those nodes where rcbuffersize is set to 512MB
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Hyperswap |
Trigger |
None |
Workaround |
Reduce rcbuffersize to a value less than 512 |
|
8.2.1.8 |
HyperSwap |
HU02077 |
All |
Critical
|
A node upgrading to v8.2.1 or later will lose access to controllers directly-attached to its FC ports and the upgrade will stall
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems that are FC direct-attached to backend storage controllers |
Trigger |
System upgrade |
Workaround |
None |
|
8.2.1.8 |
Backend Storage |
HU02086 |
All |
Critical
|
An issue, in IP Quorum, may cause a Tier 2 recovery, during initial connection to a candidate device
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2.1.0 or later that are using IP Quorum |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
IP Quorum |
HU02089 |
All |
Critical
|
Due to changes to quorum management, during an upgrade to v8.2.x, or later, there may be multiple warmstarts, with the possibility of a loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1.3 or earlier with normal configurations of more than 6 nodes, or multi-site configurations of more than 4 nodes, and no external shared MDisks |
Trigger |
Upgrading to v8.2.x or later |
Workaround |
None |
|
8.2.1.8 |
System Update |
HU02121 |
All |
Critical
|
When the system changes from copyback to rebuild a failure to clear related metadata can cause multiple node warmstarts, with the possibility of a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2.1.x, or later, that are using DRAID |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Distributed RAID |
IT26257 |
All |
Critical
|
Starting a relationship, when the remote volume is offline, may result in a T2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Hyperswap |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
HyperSwap |
IT30595 |
All |
Critical
|
A resource shortage in the RAID component can cause MDisks to be taken offline
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v8.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
RAID |
HU01836 |
All |
High Importance
|
When an auxiliary volume is moved an issue with pausing the master volume can lead to node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using Hyperswap |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
HyperSwap |
HU01942 |
FS9100, V5000, V7000 |
High Importance
|
NVMe drive ports can go offline, for a very short time, when an upgrade of that drives firmware commences
(show details)
Symptom |
None |
Environment |
FlashSystem 9100, Storwize V7000 Gen 3 and Storwize V5100 systems |
Trigger |
Start a NVMe drive firmware upgrade |
Workaround |
None |
|
8.2.1.8 |
Drives |
HU02049 |
All |
High Importance
|
GUI session handling has an issue that can generate many exceptions, adversely impacting GUI performance
(show details)
Symptom |
Performance |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Graphical User Interface |
HU02078 |
SVC |
High Importance
|
Heavily unbalanced workloads, in stretched-cluster configurations, can bias inter-node traffic through one port, adversely affecting performance
(show details)
Symptom |
Performance |
Environment |
SVC systems in a stretched-cluster configuration |
Trigger |
None |
Workaround |
Throttle or modify workloads if possible |
|
8.2.1.8 |
Performance |
HU01880 |
All |
Suggested
|
When a write, to a secondary volume, becomes stalled, a node at the primary site may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using remote copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU01936 |
All |
Suggested
|
When shrinking a volume, that has host mappings, there may be recurring node warmstarts
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.1 or later |
Trigger |
Shrink a volume while it is mapped to a host |
Workaround |
Remove all host mappings, for a volume, before performing shrinkvdisksize |
|
8.2.1.8 |
Cache |
HU02021 |
All |
Suggested
|
Disabling garbage collection may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Data Reduction Pools |
HU02085 |
All |
Suggested
|
Freeze time of Global Mirror remote copy consistency groups may not be updated correctly in certain scenarios
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.8 or later using Global Mirror |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Global Mirror |
HU02090 |
FS9100, V5000, V7000 |
Suggested
|
When a failing drive experiences an error, RAID may mishandle it, resulting in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
FlashSystem 9100 and Storwize systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
RAID |
HU02093 |
V5000 |
Suggested
|
A locking issue in the inter-node communications, of V5030 systems, can lead to a deadlock condition, resulting in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Storwize V5030 systems running v8.2.1.0 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
Reliability Availability Serviceability |
HU02099 |
All |
Suggested
|
Cloud callhome error 3201 messages may appear in the Event Log
(show details)
Symptom |
None |
Environment |
Systems running v8.2.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.8 |
System Monitoring |
IT30448 |
All |
Suggested
|
If an IP Quorum app is killed, during the commit phase of a code upgrade, then that offline IP Quorum device cannot be removed, post upgrade
(show details)
Symptom |
Configuration |
Environment |
Systems running v8.1.3, or earlier, using IP Quorum |
Trigger |
Upgrade to v8.2.0 or later, kill an IP Quorum app during the commit phase |
Workaround |
None |
|
8.2.1.8 |
IP Quorum |
IT30449 |
V5000 |
Suggested
|
Attempting to activate USB encryption on a new V5030E will fail with a CMMVCU6054E error
(show details)
Symptom |
Configuration |
Environment |
Storwize V5030E systems |
Trigger |
Attempt activation of USB encryption |
Workaround |
None |
|
8.2.1.8 |
Encryption |
HU01998 |
All |
HIPER
|
All SCSI command types can set volumes as busy resulting in I/O timeouts and multiple node warmstarts, with the possibility of an offline I/O group. For more details refer to this Flash
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Hosts |
HU02014 |
SVC |
HIPER
|
After a loss of power, where a node has a dead CMOS battery, it will fail to restart correctly. It is possible for both nodes in an I/O group to experience this issue
(show details)
Symptom |
Loss of Access to Data |
Environment |
SVC systems using SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Reliability Availability Serviceability |
HU01888 & HU01997 |
All |
Critical
|
An issue with restore mappings, in the FlashCopy component, can cause an I/O group to warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
FlashCopy |
HU01933 |
All |
Critical
|
Under rare circumstances the Data Reduction Pool deduplication rehoming process can become truncated. Subsequent detection of inconsistent metadata can lead to offline Data Reduction Pools
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1.3 or later using Deduplication |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Data Reduction Pools, Deduplication |
HU01985 |
All |
Critical
|
As a consequence of a Data Reduction Pool recovery, bad metadata may be created. When the region of disk associated with the bad metadata is accessed there may be an I/O group warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Data Reduction Pools |
HU01989 |
All |
Critical
|
For large drives, bitmap scanning, during an array rebuild, can timeout resulting in multiple node warmstarts, possibly leading to offline I/O groups
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID with drives of 8TB or more |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Distributed RAID |
HU02016 |
SVC |
Critical
|
A memory leak in the component that handles thin-provisioned MDisks can lead to an adverse performance impact with the possibility of offline MDisks. For more details refer to this Flash
(show details)
Symptom |
Offline Volumes |
Environment |
SVC systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Backend Storage |
HU02027 |
All |
Critical
|
Fabric congestion can cause internal resource constraints, in 16Gb HBAs, leading to lease expiries
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using 16Gb HBAs |
Trigger |
Fabric congestion affecting local node-node traffic |
Workaround |
Prevent fabric congestion that might affect local node-node connectivity |
|
8.2.1.6 |
Reliability Availability Serviceability |
HU02043 |
All |
Critical
|
Collecting a snap can cause nodes to run out of boot drive space and go offline with node error 565
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Taking many snap data collections on the same config node |
Workaround |
Manually delete unneeded snaps from the boot drive |
|
8.2.1.6 |
Support Data Collection |
HU02045 |
All |
Critical
|
When a node is removed from the cluster, using CLI, it may still be shown as online in the GUI. If an attempt is made to shutdown this node, from the GUI, whilst it appears to be online, then the whole cluster will shutdown
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Remove a node from the cluster using CLI. With the node showing as online, use the GUI to shut it down |
Workaround |
Manually refresh GUI browser page, after removing a node via CLI |
|
8.2.1.6 |
Graphical User Interface |
HU01890 |
All |
High Importance
|
FlashCopy mappings, from master volume to primary change volume, may become stalled when a T2 recovery occurs whilst the mappings are in a copying state
(show details)
Symptom |
None |
Environment |
Systems using Global Mirror with Change Volumes |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Global Mirror With Change Volumes |
HU02037 |
All |
High Importance
|
A FlashCopy consistency group, with a mix of mappings in different states, cannot be stopped
(show details)
Symptom |
None |
Environment |
Systems using FlashCopy |
Trigger |
Some, but not all, mappings, in a consistency group, have their target volumes run out of space |
Workaround |
None |
|
8.2.1.6 |
FlashCopy |
HU02053 |
FS9100, V5100, V7000 |
High Importance
|
An issue with canister BIOS update can stall system upgrades
(show details)
Symptom |
Loss of Redundancy |
Environment |
FS9100, V7000 Gen 3 and V5100 systems |
Trigger |
Upgrade to v8.3.0 |
Workaround |
None |
|
8.2.1.6 |
System Update |
HU02055 |
All |
High Importance
|
Creating a FlashCopy snapshot, in the GUI, does not set the same preferred node for both source and target volumes. This may adversely impact performance
(show details)
Symptom |
Performance |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
Use the movevdisk command to manually set the same preferred node for both the source and target volumes in the FC map |
|
8.2.1.6 |
FlashCopy |
HU02072 |
All |
High Importance
|
An issue in the handling of email transmission can write a large file to the node boot drive. If this causes the boot drive to become full, the node will go offline with error 565
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v8.2.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
System Monitoring |
HU01843 |
All |
Suggested
|
A node hardware issue can cause a CLI command to timeout resulting in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Command Line Interface |
HU01892 |
All |
Suggested
|
LUNs of greater than 2TB, presented by HP XP7 storage controllers, are not supported
(show details)
Symptom |
Configuration |
Environment |
Systems with HP XP7 backend controllers |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Backend Storage |
HU01974 |
All |
Suggested
|
With all Remote Support Assistant connections closed, the GUI may show that a connection is still in progress
(show details)
Symptom |
None |
Environment |
Systems running v8.1 or later using Remote Support Assistance |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
System Monitoring |
HU01978 |
All |
Suggested
|
Unable to create HyperSwap volumes. The mkvolume command fails with CMMVC7050E error
(show details)
Symptom |
None |
Environment |
Systems running v8.2 or later using HyperSwap |
Trigger |
None |
Workaround |
Use the early (pre mkvolume) procedure for creating HyperSwap volumes |
|
8.2.1.6 |
HyperSwap |
HU01979 |
All |
Suggested
|
The figure for used_virtualization, in the output of a lslicense command, may be unexpectedly large
(show details)
Symptom |
None |
Environment |
Systems running v8.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Command Line Interface |
HU01982 |
All |
Suggested
|
In an environment, with multiple IP Quorum servers, if the quorum component encounters a duplicate UID then a node may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.2.1 or later running IP quorum |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
IP Quorum |
HU01983 |
All |
Suggested
|
Improve debug data capture to assist in determining the reason for a Data Reduction Pool to be taken offline
(show details)
Symptom |
None |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Data Reduction Pools |
HU01986 |
All |
Suggested
|
An accounting issue in the FlashCopy component may cause node warmstarts
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
FlashCopy |
HU01991 |
All |
Suggested
|
An issue in the handling of extent allocation, in the Data Reduction Pool component, can cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Data Reduction Pools |
HU02020 |
FS9100, V5000, V7000 |
Suggested
|
An internal hardware bus, running at the incorrect speed, may give rise to spurious DIMM over-temperature errors
(show details)
Symptom |
None |
Environment |
FlashSystem 9100, Storwize V7000 Gen 3 and Storwize V5100 systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Reliability Availability Serviceability |
HU02029 |
All |
Suggested
|
An issue with the SSMTP process may result in failed callhome, inventory reporting and user notifications. A testemail command will fail with a CMMVC9051E error
(show details)
Symptom |
None |
Environment |
Systems running v8.2.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
System Monitoring |
HU02039 |
All |
Suggested
|
An issue in the management steps of Data Reduction Pool recovery may lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Data Reduction Pools |
HU02067 |
All |
Suggested
|
If multiple recipients are specified, for callhome emails, then no callhome emails will be sent
(show details)
Symptom |
None |
Environment |
Systems running v8.2.1.5 or later |
Trigger |
Specify multiple recipients for callhome email messages |
Workaround |
None |
|
8.2.1.6 |
System Monitoring |
HU02129 |
All |
Suggested
|
GUI drive filtering fails with An error occurred loading table data
(show details)
Symptom |
None |
Environment |
Systems running v8.2.1 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.6 |
Graphical User Interface |
HU02007 |
All |
HIPER
|
During volume migration an issue, in the handling of old to new extents transfer, can lead to cluster-wide warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.5 |
Storage Virtualisation |
HU02009 |
All |
Critical
|
Systems which are using Data Reduction Pools, with the maximum possible extent size of 8GB, and which experience a very specific I/O workload, may experience an issue due to garbage collection. This can cause repeated node warmstarts and loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.5 |
Data Reduction Pools |
HU02011 |
All |
High Importance
|
When a node warmstart occurs on a system using Data Reduction Pools, there is a small possibility that the node will not automatically return online. If the partner node is also offline, this can cause temporary loss of access to data
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.5 |
Data Reduction Pools |
HU02012 |
All |
High Importance
|
Under certain I/O workloads the garbage collection process can adversely impact volume write response times
(show details)
Symptom |
Performance |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.5 |
Data Reduction Pools |
HU01865 |
All |
HIPER
|
When creating a HyperSwap relationship, using addvolumecopy (or similar methods), the system should perform a synchronisation operation to copy the data from the original copy to the new copy. In some rare cases this synchronisation is skipped, leaving the new copy with bad data (all zeros)
(show details)
Symptom |
Data Integrity Loss |
Environment |
Systems running v7.5 or later using HyperSwap |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
HyperSwap |
HU01918 |
All |
HIPER
|
Where Data Reduction Pools have been created on earlier code levels, upgrading the system, to an affected release, can cause an increase in the level of concurrent flushing to disk. This may result in a loss of access to data. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1.3.4, v8.2.0.3 or v8.2.1.x using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Data Reduction Pools |
HU02008 |
All |
HIPER
|
When a DRAID rebuild occurs, occasionally a RAID deadlock condition can be triggered by a particular type of I/O workload. This can lead to repeated node warmstarts and a loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Distributed RAID |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Distributed RAID |
HU01887 |
All |
Critical
|
In circumstances where host configuration data becomes inconsistent, across nodes, an issue in the CLI policing code may cause multiple warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Host Clusters |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Command Line Interface, Host Cluster |
HU01900 |
All |
Critical
|
Executing a command, that can result in a shared mapping being created or destroyed, for an individual host, in a host cluster, without that command applying to all hosts in the host cluster, may lead to multiple node warmstarts with the possibility of a T2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Execute a command that can result in a shared mapping being created or destroyed |
Workaround |
Do not use the -force option when running a rmhostiogrp command |
|
8.2.1.4 |
Host Cluster |
HU01910 |
All |
Critical
|
When FlashCopy mappings are created, with a grain size of 64KB, it is possible for an overflow condition in the bitmap to occur. This can resulting in multiple node warmstarts with a possible loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using FlashCopy mappings with a 64KB grain size |
Trigger |
None |
Workaround |
Select a grain size of 256KB when creating FlashCopy mappings |
|
8.2.1.4 |
FlashCopy |
HU01928 |
All |
Critical
|
When two IOs attempt to access the same address, the state of the data may be incorrectly set to invalid causing offline volumes and, possibly, offline pools
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Data Reduction Pools |
HU01987 |
SVC |
Critical
|
During SAN fabric power maintenance a cluster may breech resource limits, on the remaining node to node links, resulting in system-wide lease expiry
(show details)
Symptom |
Loss of Access to Data |
Environment |
SVC model SV1 systems running v8.2.1 or later |
Trigger |
Remove a connected SAN switch |
Workaround |
If only two local ports are in use due to zoning and masking for remote copy, and one fabric needs maintenance: Stop partnership; Change port masking - allow node to node comms to use the ports that were remote ports; Perform SAN maintenance |
|
8.2.1.4 |
Reliability Availability Serviceability |
HU02000 |
All |
Critical
|
Data Reduction Pools may go offline due to a timing issue in metadata handling
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Data Reduction Pools |
HU02013 |
All |
Critical
|
A race condition between the extent invalidation and destruction in the garbage collection process may cause a node warmstart with the possibility of offline volumes
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Data Reduction Pools |
HU02025 |
All |
Critical
|
An issue with metadata handling, where a pool has been taken offline, may lead to an out of space condition in that pool preventing its return to operation
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Data Reduction Pools |
HU01886 |
All |
High Importance
|
The Unmap function can leave volume extents, that have not been freed, preventing managed disk and pool removal
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
SCSI Unmap |
HU01902 |
V5000, V7000 |
High Importance
|
During an upgrade, an issue with VPD migration, can cause a timeout leading to a stalled upgrade
(show details)
Symptom |
Loss of Redundancy |
Environment |
Storwize systems |
Trigger |
Upgrade |
Workaround |
None |
|
8.2.1.4 |
System Update |
HU01925 |
FS9100 |
High Importance
|
Systems will incorrectly report offline and unresponsive NVMe drives after an I/O group outage. These errors will fail to auto-fix and must be manually marked as fixed
(show details)
Symptom |
None |
Environment |
FlashSystem 9100 family systems |
Trigger |
Both canisters in I/O group go offline |
Workaround |
None |
|
8.2.1.4 |
System Monitoring |
HU01930 |
FS9100 |
High Importance
|
Certain types of FlashCore Module (FCM) failure may not result in a call home, delaying the shipment of a replacement
(show details)
Symptom |
None |
Environment |
FlashSystem 9100 family systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Drives |
HU01937 |
FS9100, V7000 |
High Importance
|
DRAID copy-back operation can overload NVMe drives resulting in high I/O latency
(show details)
Symptom |
Performance |
Environment |
Systems with NVMe drives using DRAID |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Distributed RAID, Drives |
HU01939 |
FS9100, V7000 |
High Importance
|
After replacing a canister, and attempting to bring the new canister into the cluster, it may remain offline
(show details)
Symptom |
Loss of Redundancy |
Environment |
FlashSystem 9100 family and Storwize V7000 Gen 3 systems running v8.2.1.0 or later |
Trigger |
Using the 'satask chbootdrive -replacecanister' command during node replacement |
Workaround |
Replace both canister boot drives at the same time with new drives |
|
8.2.1.4 |
Reliability Availability Serviceability |
HU01941 |
All |
High Importance
|
After upgrading the system to v8.2, or later, when expanding a mirrored volume, the formatting of additional space may become stalled
(show details)
Symptom |
None |
Environment |
Systems upgrading from v8.1.3 or earlier to v8.2.0 or later |
Trigger |
Single or mirrored copy volume exists on system while running v8.1.3.x or earlier Cluster is upgraded from 813 or earlier to 820 or later No extent migrations or changes in host mapping for volume Add copy to volume to make it a volume mirror No extent migrations or changes in host mapping for volume Expand volume size Original volume will not progress format whilst copy will complete Also where both copies of the mirrored volume existed, before the upgrade from v8.1.3.x or earlier, both can have stuck formatting |
Workaround |
None |
|
8.2.1.4 |
Volume Mirroring |
HU01944 |
All |
High Importance
|
Proactive host failover not waiting for 25 seconds before allowing nodes to go offline during upgrades or maintenance
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Reliability Availability Serviceability |
HU01945 |
All |
High Importance
|
Systems with Flash Core Modules are unable to upgrade the firmware for those drives
(show details)
Symptom |
None |
Environment |
Systems using IBM Flash Core Modules |
Trigger |
FCM firmware upgrade |
Workaround |
None |
|
8.2.1.4 |
Drives |
HU01971 |
FS9100, V7000 |
High Importance
|
Spurious DIMM over-temperature errors may cause a node to go offline with node error 528
(show details)
Symptom |
Loss of Redundancy |
Environment |
FlashSystem 9100 family and Storwize V7000 Gen 3 systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Reliability Availability Serviceability |
HU01972 |
All |
High Importance
|
When an array is in a quiescing state, for example where a member has been deleted, I/O may become pended leading to multiple warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
All systems |
Trigger |
Delete an array member using "charraymember -used unused" command |
Workaround |
None |
|
8.2.1.4 |
Distributed RAID, RAID |
HU00744 |
All |
Suggested
|
Single node warmstart due to an accounting issue within the cache component
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.3 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Cache |
HU01485 |
SVC |
Suggested
|
When a SV1 node is started, with only one PSU powered, powering up the other PSU will not extinguish the Power Fault LED.Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed
(show details)
Symptom |
None |
Environment |
SVC systems using SV1 model nodes |
Trigger |
Power up node with only one PSU powered. Power Fault LED is lit. Power up other PSU. Power Fault LED remains lit. |
Workaround |
Ensure both PSUs are powered before starting node |
|
8.2.1.4 |
System Monitoring |
HU01659 |
SVC |
Suggested
|
Node Fault LED can be seen to flash in the absence of an error condition.Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed
(show details)
Symptom |
None |
Environment |
SVC systems using SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
System Monitoring |
HU01857 |
All |
Suggested
|
Improved validation of user input in GUI
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Graphical User Interface |
HU01860 |
All |
Suggested
|
During garbage collection the flushing of extents may become stuck leading to a timeout and a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Data Reduction Pools |
HU01869 |
All |
Suggested
|
Volume copy deletion, in a Data Reduction Pool, triggered by rmvdiskcopy rmvolumecopy or addvdiskcopy -autodelete (or similar) may become stalled with the copy being left in deleting status
(show details)
Symptom |
None |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Data Reduction Pools |
HU01911 |
All |
Suggested
|
The System Overview screen, in the GUI, may display nodes in the wrong site
(show details)
Symptom |
None |
Environment |
Systems in a stretched cluster or HyperSwap topology |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Graphical User Interface |
HU01912 |
All |
Suggested
|
Systems with iSCSI-attached controllers may see node warmstarts due to I/O request timeouts
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.2.1.0 or later with iSCSI-attached controllers |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Backend Storage |
HU01915 & IT28654 |
All |
Suggested
|
Systems, with encryption enabled, that are using key servers to manage encryption keys, may fail to connect to the key servers if the servers SSL certificates are part of a chain of trust
(show details)
Symptom |
None |
Environment |
Systems with encryption enabled |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Encryption |
HU01916 |
All |
Suggested
|
The GUI Dashboard and the CLI lssystem command report physical capacity incorrectly
(show details)
Symptom |
None |
Environment |
Systems running v8.1 or later |
Trigger |
Upgrading from v8.1 or later |
Workaround |
lsmdisk can continue to be used to provide accurate reporting |
|
8.2.1.4 |
Command Line Interface, Graphical User Interface |
HU01926 |
SVC, V7000 |
Suggested
|
When a node, with 32GB of RAM, is upgraded to v8.2.1 it may experience a warmstart resulting in a failed upgrade
(show details)
Symptom |
None |
Environment |
Systems with nodes that have 32GB RAM |
Trigger |
Upgrade to v8.2.1 |
Workaround |
None |
|
8.2.1.4 |
System Update |
HU01929 |
FS9100, V7000 |
Suggested
|
Drive fault type 3 (error code 1686) may be seen in the Event Log for empty slots
(show details)
Symptom |
None |
Environment |
FlashSystem 9100 family and Storwize V7000 Gen 3 systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
System Monitoring |
HU01959 |
All |
Suggested
|
An timing window issue in the Thin Provisioning component can cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using FlashCopy and Thin Provisioning |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
FlashCopy, Thin Provisioning |
HU01961 |
V5000, V7000 |
Suggested
|
A hardware issue can provoke the system to repeatedly try to collect a statesave, from the enclosure management firmware, causing 1048 errors in the Event Log
(show details)
Symptom |
None |
Environment |
Storwize systems |
Trigger |
None |
Workaround |
Replace the canister |
|
8.2.1.4 |
System Monitoring |
HU01962 |
All |
Suggested
|
When Call Home servers return an invalid message it can be incorrectly reported as an error 3201 in the Event Log
(show details)
Symptom |
None |
Environment |
Systems with Call Home configured |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
System Monitoring |
HU01976 |
All |
Suggested
|
A new MDisk array may not be encrypted even though encryption is enabled on the system
(show details)
Symptom |
None |
Environment |
Systems running v8.2 or later using encryption without an I/O Group id of 0 |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Encryption |
HU02001 |
All |
Suggested
|
During a system upgrade an issue in callhome may cause a node warmstart stalling the upgrade
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v8.1 or earlier |
Trigger |
Upgrade system to v8.2 or later |
Workaround |
Before a system upgrade: Remove email servers; Disable cloud call home |
|
8.2.1.4 |
System Monitoring |
HU02002 |
All |
Suggested
|
On busy systems, diagnostic data collection may not complete correctly producing livedumps with missing pages
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Support Data Collection |
HU02019 |
All |
Suggested
|
When the master and auxiliary volumes, in a relationship, have the same name it is not possible, in the GUI, to determine which is master or auxiliary
(show details)
Symptom |
None |
Environment |
Systems using remote copy |
Trigger |
Identical naming of related master and auxiliary volumes |
Workaround |
Use distinguishable names for remote copy master and auxiliary volumes |
|
8.2.1.4 |
Graphical User Interface |
HU02166 |
All |
Suggested
|
A timing window issue, in RAID code that handles recovery after a drive has been taken out of sync, due to a slow I/O, can cause a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
RAID |
IT28433 |
All |
Suggested
|
Timing window issue in the Data Reduction Pool rehoming component can cause a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
Data Reduction Pools |
IT28728 |
All |
Suggested
|
Email alerts will not work where the mail server does not allow unqualified client host names
(show details)
Symptom |
Configuration |
Environment |
Systems running v8.2 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.4 |
System Monitoring |
HU01932 |
All |
Critical
|
When a rmvdisk command initiates a Data Reduction Pool rehoming process any I/O to the removed volume may cause multiple warmstarts leading to a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.2.1.1 using Deduplication |
Trigger |
rmvdisk command will initiate a Data Reduction Pool rehoming process |
Workaround |
None |
|
8.2.1.2 |
Deduplication |
HU01920 |
All |
Critical
|
An issue in the garbage collection process can cause node warmstarts and offline pools
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.1 |
Data Reduction Pools |
HU01492 & HU02024 |
SVC, V5000, V7000 |
HIPER
|
All ports of a 16Gb HBA can be affected when a single port is congested. This can lead to lease expiries if all ports used for inter-node communication are on the same FC adapter
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using 16Gb HBAs |
Trigger |
All ports used for inter-node communication are on the same FC adapter and a port on that adapter experiences congestion |
Workaround |
Separate inter-node traffic so that multiple adapters are used |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01617 |
All |
HIPER
|
Due to a timing window issue, stopping a FlashCopy mapping, with the -autodelete option, may result in a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
Avoid stopping FlashCopy mappings with the -autodelete option |
|
8.2.1.0 |
FlashCopy |
HU01828 |
All |
HIPER
|
Node warmstarts may occur during deletion of deduplicated volumes due to a timing-related issue
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using deduplicated volume copies |
Trigger |
Deleting a deduplication volume copy |
Workaround |
Do not delete deduplicated volume copies |
|
8.2.1.0 |
Deduplication |
HU01851 |
All |
HIPER
|
When a deduplicated volume is deleted there may be multiple node warmstarts and offline pools
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1.3 or later using Deduplication |
Trigger |
Delete a deduplicated volume |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools, Deduplication |
HU01873 |
All |
HIPER
|
Deleting a volume, in a Data Reduction Pool, while volume protection is enabled and when the volume was not explicitly unmapped, before deletion, may result in simultaneous node warmstarts. For more details refer to this Flash
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools |
Trigger |
Delete volume in Data Reduction Pool while volume protection is enabled |
Workaround |
Either: Disable volume protection; or Remove host mappings before deleting a volume. If using scripts, modify them to unmap volumes before deletion. |
|
8.2.1.0 |
Data Reduction Pools |
HU01906 |
FS9100 |
HIPER
|
Low-level hardware errors may not be recovered correctly, causing a canister to reboot. If multiple canisters reboot, this may result in loss of access to data
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
FlashSystem 9100 family systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01913 |
All |
HIPER
|
A timing window issue in the DRAID6 rebuild process can cause node warmstarts with the possibility of a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Distributed RAID |
HU01743 |
All |
Critical
|
Where hosts are directly attached a mishandling of the login process, by the fabric controller, may result in dual node warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems with direct-attached hosts |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Hosts |
HU01758 |
All |
Critical
|
After an unexpected power loss, all nodes, in a cluster, may warmstart repeatedly, necessitating a Tier 3 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
Power outage |
Workaround |
None |
|
8.2.1.0 |
RAID |
HU01799 |
All |
Critical
|
Timing window issue can affect operation of the HyperSwap addvolumecopy command causing all nodes to warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using HyperSwap |
Trigger |
None |
Workaround |
Ensuring that the volume is not in a FlashCopy mapping will allow the use of the addvolumecopy. Alternatively the addvolumecopy command combines a number of CLI commands for convenience. Use these CLI commands individually to achieve the same outcome |
|
8.2.1.0 |
HyperSwap |
HU01825 |
All |
Critical
|
Invoking a chrcrelationship command when one of the relationships in a consistency group is running in the opposite direction to the others may cause a node warmstart followed by a Tier 2 recovery
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using HyperSwap |
Trigger |
A relationship running in one direction is added to a consistency group running in the other direction whilst one of the FlashCopy maps associated with the HyperSwap relationship is still stopping/cleaning |
Workaround |
Do not add a relationship to a consistency group if they are running in opposite directions (i.e. the Primary of the consistency group and the Primary of the relationship are on different sites); Do not add a relationship to a consistency group if the relationship still has one of its FlashCopy maps in the stopping state. The clean progress needs to reach 100 percent before the relationship can be safely added. |
|
8.2.1.0 |
FlashCopy |
HU01833 |
All |
Critical
|
If both nodes, in an I/O group, start up together a timing window issue may occur, that would prevent them running garbage collection, leading to a related Data Reduction Pool running out of space
(show details)
Symptom |
Offline Volumes |
Environment |
Systems using Data Reduction Pools |
Trigger |
Start both nodes in an I/O group at the same time |
Workaround |
Ensure nodes in an I/O group start one at a time |
|
8.2.1.0 |
Data Reduction Pools |
HU01845 |
All |
Critical
|
If the execution of a rmvdisk -force command, for the FlashCopy target volume in a GMCV relationship, coincides with the start of a GMCV cycle all nodes may warmstart
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.7.1 or later using Global Mirror with Change Volumes |
Trigger |
rmvdisk -force command execution coincides with the start of a GMCV cycle |
Workaround |
Remove the related FC map before issuing a rmvdisk command without -force |
|
8.2.1.0 |
Global Mirror With Change Volumes |
HU01847 |
All |
Critical
|
FlashCopy handling of medium errors across a number of drives on backend controllers may lead to multiple node warmstarts
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v7.8.1 or later using FlashCopy |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
FlashCopy |
HU01848 |
All |
Critical
|
During an upgrade, systems with a large AIX VIOS setup may have multiple node warmstarts with the possibility of a loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems presenting storage to large IBM AIX VIOS configurations |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
System Update |
HU01850 |
All |
Critical
|
When the last deduplication-enabled volume copy in a Data Reduction Pool is deleted the pool may go offline temporarily
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools with deduplicated volume copies |
Trigger |
Delete last deduplication-enabled volume copy in a Data Reduction Pool |
Workaround |
If a Data Reduction Pool contains volumes with deduplication enabled keep at least one of those volumes in the pool |
|
8.2.1.0 |
Data Reduction Pools, Deduplication |
HU01855 |
All |
Critical
|
Clusters using Data Reduction Pools can experience multiple warmstarts on all nodes putting them in a service state
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems running v8.1.2 or later using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools |
HU01862 |
All |
Critical
|
When a Data Reduction Pool is removed, and the -force option is specified, there may be a temporary loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools |
Trigger |
Remove a Data Reduction Pool with the -force option |
Workaround |
Do not use -force option when removing a Data Reduction Pool |
|
8.2.1.0 |
Data Reduction Pools |
HU01876 |
All |
Critical
|
Where systems are connected to controllers, that have FC ports that are capable of acting as initiators and targets, when NPIV is enabled then node warmstarts can occur
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems, with NPIV enabled, attached to host ports that can act as SCSI initiators and targets |
Trigger |
Zone host initiator and target ports in with the target port WWPN then enable NPIV |
Workaround |
Unzone host or disable NPIV |
|
8.2.1.0 |
Backend Storage |
HU01878 |
All |
Critical
|
During an upgrade from v7.8.1 or earlier to v8.1.3 or later if an MDisk goes offline then at completion all volumes may go offline
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v7.8.1 or earlier |
Trigger |
MDisk goes offline during an upgrade to v8.1.3 or later |
Workaround |
None |
|
8.2.1.0 |
System Update |
HU01885 |
All |
Critical
|
As writes are made to a Data Reduction Pool it is necessary to allocate new physical capacity. Under unusual circumstances it is possible for the handling of an expansion request to stall further I/O leading to node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools |
HU01901 |
V7000 |
Critical
|
Enclosure management firmware, in an expansion enclosure, will reset a canister after a certain number of discovery requests have been received, from the controller, for that canister. It is possible simultaneous resets may occur in adjacent canisters causing a temporary loss of access to data
(show details)
Symptom |
Loss of Access to Data |
Environment |
Storwize expansion enclosures running v8.2.0 |
Trigger |
One million discovery requests received from attached controller |
Workaround |
Proactively reset individual expansion canisters at convenient times |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01957 |
All |
Critical
|
Due to an issue in Data Reduction Pools, when the system attempts an upgrade, there may be node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems using Data Reduction Pools |
Trigger |
Initiate system upgrade |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools, System Update |
HU01965 |
All |
Critical
|
A timing window issue in the deduplication component can lead to I/O timeouts, and a node warmstart, with the possibility of an offline MDisk group
(show details)
Symptom |
Offline Volumes |
Environment |
Systems running v8.1.3 or later using Deduplication |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Deduplication |
HU02042 |
All |
Critical
|
An issue in the handling of metadata, after a Data Reduction Pool recovery operation, can lead to repeated node warmstarts, putting an I/O group into a service state
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using Data Reduction Pools |
Trigger |
T3 recovery |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools |
IT25850 |
All |
Critical
|
I/O performance may be adversely affected towards the end of DRAID rebuilds. For some systems there may be multiple warmstarts leading to a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Distributed RAID |
IT27460 |
All |
Critical
|
Lease expiry can occur between local nodes when remote connection is lost, due to the mishandling of messaging credits
(show details)
Symptom |
Loss of Access to Data |
Environment |
All systems |
Trigger |
None |
Workaround |
Use four ports for local to local node communications, on at least two separate fibre channel adapters per node. Port mask so that all four are usable. Use a different fibre channel adapter than the above two adapters for remote port communications. If there are issues with the FCIP tunnel, temporarily block that until it is fixed. |
|
8.2.1.0 |
Reliability Availability Serviceability |
IT29040 |
All |
Critical
|
Occasionally a DRAID rebuild, with drives of 8TB or more, can encounter an issue which causes node warmstarts and potential loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Systems using DRAID with drives of 8TB or more |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Distributed RAID, RAID |
IT29853 |
V5000 |
Critical
|
After upgrading to v8.1.1, or later, V5000 Gen 2 systems, with Gen 1 expansion enclosures, may experience multiple node warmstarts leading to a loss of access
(show details)
Symptom |
Loss of Access to Data |
Environment |
Storwize V5000 Gen 2 systems with Gen 1 expansion enclosures |
Trigger |
Upgrade to v8.1.1 or later |
Workaround |
None |
|
8.2.1.0 |
System Update |
HU01507 |
All |
High Importance
|
Until the initial synchronisation process completes, high system latency may be experienced when a volume is created with two compressed copies or when space-efficient copy is added to a volume with an existing compressed copy
(show details)
Symptom |
Performance |
Environment |
All systems |
Trigger |
Create a volume with two compressed copies or add a space-efficient copy to a volume with an existing compressed copy |
Workaround |
Avoid: creating a new volume with two compressed copies; adding a SE volume copy to a volume that already possesses a compressed copy |
|
8.2.1.0 |
Volume Mirroring |
HU01661 |
All |
High Importance
|
A cache-protection mechanism flag setting can become stuck leading to repeated stops of consistency group synchronisation
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v7.6 or later using remote copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
HyperSwap |
HU01733 |
All |
High Importance
|
Canister information, for the High Density Expansion Enclosure, may be incorrectly reported
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems using the High Density Expansion Enclosure (92F) |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01761 |
All |
High Importance
|
Entering multiple addmdisk commands, in rapid succession, to more than one storage pool, may cause node warmstarts
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v8.1 or later with two or more storage pools |
Trigger |
Run multiple addmdisk commands to more than one storage pool at the same time |
Workaround |
Paced addmdisk commands to one storage pool at a time |
|
8.2.1.0 |
Backend Storage |
HU01797 |
All |
High Importance
|
Hitachi G1500 backend controllers may exhibit higher than expected latency
(show details)
Symptom |
Performance |
Environment |
Systems with Hitachi G1500 backend controllers |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Backend Storage |
HU01810 |
All |
High Importance
|
Deleting volumes, or using FlashCopy/Global Mirror with Change Volumes, in a Data Reduction Pool, may impact the performance of other volumes in the pool
(show details)
Symptom |
Performance |
Environment |
Systems using Data Reduction Pools |
Trigger |
Volume deletion or FlashCopy/GMCV |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools |
HU01837 |
All |
High Importance
|
In systems where a vVols metadata volume has been created an upgrade to v8.1.3 or later will cause a node warmstart stalling the upgrade
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems running v8.1.0, v8.1.1 or v8.1.2 that are providing vVols |
Trigger |
Upgrading to v8.1.3 or later |
Workaround |
Contact support if system is running v8.1.2. Otherwise this workaround can be used: Use svcinfo lsmetadatavdisk to find the volume id; Create a new volume copy in the same MDisk group - svctask addvdiskcopy -mdiskgrp X -autodelete <vdisk_id>; Wait until lsvdisksyncprogress no longer shows a mirror in progress; Upgrade |
|
8.2.1.0 |
System Update, vVols |
HU01839 |
All |
High Importance
|
Where a VMware host is being served volumes, from two different controllers, and an issue, on one controller, causes the related volumes to be taken offline then I/O performance, for the volumes from the other controller, will be adversely affected
(show details)
Symptom |
Performance |
Environment |
Systems running v7.5 or later presenting volumes to VMware hosts, from more than one back-end controller |
Trigger |
Issue on back-end controller takes volumes offline |
Workaround |
None |
|
8.2.1.0 |
Hosts |
HU01842 |
All |
High Importance
|
Bursts of I/O to Read-Intensive Drives can be interpreted as dropped frames against the resident slots, leading to redundant drives being incorrectly failed
(show details)
Symptom |
Loss of Redundancy |
Environment |
Systems with Read-Intensive Drives |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Drives |
HU01846 |
SVC |
High Importance
|
Silent battery discharge condition will unexpectedly take a SVC node offline putting it into a 572 service state
(show details)
Symptom |
Loss of Redundancy |
Environment |
SVC systems using DH8 & SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01852 |
All |
High Importance
|
The garbage collection rate can lead to Data Reduction Pools running out of space even though reclaimable capacity is available
(show details)
Symptom |
None |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools |
HU01858 |
All |
High Importance
|
Total used capacity of a Data Reduction Pool within a single I/O group is limited to 256TB. Garbage collection does not correctly recognise this limit. This may lead to a pool running out of free capacity and going offline
(show details)
Symptom |
None |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools |
HU01881 |
FS9100 |
High Importance
|
An issue within the compression card in FS9100 systems can result in the card being incorrectly flagged as failed leading to warmstarts
(show details)
Symptom |
Loss of Redundancy |
Environment |
FS9100 systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Compression |
HU01883 |
All |
High Importance
|
Config node processes may consume all available memory, leading to node warmstarts. This can be caused, for example, by large numbers of concurrent SSH connections being opened
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
Systems running v8.1.3 or later |
Trigger |
Many concurrent SSH connections |
Workaround |
Reduce number of SSH connections |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01907 |
SVC |
High Importance
|
An issue in the handling of the power cable sense registers can cause a node to be put into service state with a 560 error
(show details)
Symptom |
Loss of Redundancy |
Environment |
SVC systems using SV1 model nodes |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01934 |
FS9100 |
High Importance
|
An issue in the handling of faulty canister components can lead to multiple node warmstarts for that canister
(show details)
Symptom |
Multiple Node Warmstarts |
Environment |
FlashSystem 9100 family systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU00921 |
All |
Suggested
|
A node warmstart may occur when an MDisk state change gives rise to duplicate discovery processes
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
|
HU01276 |
All |
Suggested
|
An issue in the handling of debug data from the FC adapter can cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01523 |
All |
Suggested
|
An issue with FC adapter initialisation can lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01564 |
All |
Suggested
|
FlashCopy maps cleaning process is not monitoring the grains correctly which may cause FlashCopy maps to not stop
(show details)
Symptom |
None |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
FlashCopy |
HU01571 |
All |
Suggested
|
An upgrade can become stalled due to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems undergoing a code upgrade |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
System Update |
HU01657 |
SVC, V5000, V7000 |
Suggested
|
The 16Gb FC HBA firmware may experience an issue, with the detection of unresponsive links, leading to a single node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01667 |
All |
Suggested
|
A timing-window issue, in the remote copy component, may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using remote copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU01719 |
All |
Suggested
|
Node warmstart due to a parity error in the HBA driver firmware
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems running v7.6 and later using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01737 |
All |
Suggested
|
On the Update System screen, for Test Only, if a valid code image is selected, in the Run Update Test Utility dialog, then clicking the Test button will initiate a system update
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
Select a valid code image in the "Run Update Test Utility" dialog and click "Test" button |
Workaround |
Do not select a valid code image in the "Test utility" field of the "Run Update Test Utility" dialog |
|
8.2.1.0 |
System Update |
HU01760 |
All |
Suggested
|
FlashCopy map progress appears to be stuck at zero percent
(show details)
Symptom |
None |
Environment |
Systems using FlashCopy |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
FlashCopy |
HU01765 |
All |
Suggested
|
Node warmstart may occur when there is a delay to I/O at the secondary site
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using remote copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU01772 |
All |
Suggested
|
The mail queue may become blocked preventing the transmission of event log messages
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
System Monitoring |
HU01784 |
All |
Suggested
|
If a cluster using IP quorum experiences a site outage, the IP quorum device may become invalid. Restarting the quorum application will resolve the issue
(show details)
Symptom |
Configuration |
Environment |
Systems using Hyperswap or enhanced stretched cluster |
Trigger |
Outage on all nodes at a single site |
Workaround |
Restart the quorum application after the site outage |
|
8.2.1.0 |
HyperSwap, Quorum |
HU01786 |
All |
Suggested
|
An issue in the monitoring of SSD write endurance can result in false 1215/2560 errors in the Event Log
(show details)
Symptom |
None |
Environment |
Systems running v7.7.1 or later with SSDs |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Drives |
HU01791 |
All |
Suggested
|
Using the chhost command will remove stored CHAP secrets
(show details)
Symptom |
Configuration |
Environment |
Systems using iSCSI |
Trigger |
Run the "chhost -gui -name <host name> <host id>" command after configuring CHAP secret |
Workaround |
Set the CHAP secret whenever changing the host name |
|
8.2.1.0 |
iSCSI |
HU01807 |
All |
Suggested
|
The lsfabric command may show incorrect local node id and local node name for some Fibre Channel logins
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
None |
Workaround |
Use the local WWPN and reference the node in lsportfc to get the correct information |
|
8.2.1.0 |
Command Line Interface |
HU01811 |
All |
Suggested
|
DRAID rebuilds, for large (>10TB) drives, may require lengthy metadata processing leading to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using DRAID |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Distributed RAID |
HU01815 |
All |
Suggested
|
In Data Reduction Pools, volume size is limited to 96TB
(show details)
Symptom |
None |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools |
HU01817 |
All |
Suggested
|
Volumes used for vVols metadata or cloud backup, that are associated with a FlashCopy mapping, cannot be included in any further FlashCopy mappings
(show details)
Symptom |
Configuration |
Environment |
Systems using vVols or TCT |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
FlashCopy |
HU01821 |
SVC |
Suggested
|
An attempt to upgrade a two-node enhanced stretched cluster fails due to incorrect volume dependencies
(show details)
Symptom |
None |
Environment |
Systems configured as a two-node enhanced stretched cluster that are using Data Reduction Pools |
Trigger |
Upgrade |
Workaround |
Revert cluster to standard topology and remove site settings from nodes and controllers for the duration of the upgrade |
|
8.2.1.0 |
Data Reduction Pools, System Update |
HU01832 |
All |
Suggested
|
Creation and distribution of the config file may cause an out-of-memory condition, leading to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU01849 |
All |
Suggested
|
An excessive number of SSH sessions may lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
Initiate a large number of SSH sessions (e.g. one session every 5 seconds) |
Workaround |
Avoid initiating excessive numbers of SSH sessions |
|
8.2.1.0 |
System Monitoring |
HU01856 |
All |
Suggested
|
A garbage collection process can time out waiting for an event in the partner node resulting in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools |
HU01863 |
All |
Suggested
|
In rare circumstances, a drive replacement may result in a ghost drive (i.e. a drive with the same ID as the replaced drive stuck in a permanently offline state)
(show details)
Symptom |
None |
Environment |
All systems |
Trigger |
Drive replacement |
Workaround |
None |
|
8.2.1.0 |
Drives |
HU01871 |
All |
Suggested
|
An issue with bitmap synchronisation can lead to a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Data Reduction Pools |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Data Reduction Pools |
HU01879 |
All |
Suggested
|
Latency induced by DWDM inter-site links may result in a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using DWDM inter-site links |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
|
HU01893 |
FS9100, SVC, V7000 |
Suggested
|
Excessive reporting frequency of NVMe drive diagnostics generates large numbers of callhome events
(show details)
Symptom |
None |
Environment |
Systems running v8.2 or later using NVMe drives |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Drives |
HU01895 |
All |
Suggested
|
Where a banner has been created, without a new line at the end, any subsequent T4 recovery will fail
(show details)
Symptom |
None |
Environment |
Systems using DRAID with a banner set |
Trigger |
Initiate a T4 recovery |
Workaround |
Modify svc.config.backup.xml - add new line at the end of banner stanza. Before the </object> tag |
|
8.2.1.0 |
Distributed RAID |
HU01981 |
All |
Suggested
|
Although an issue, in the HBA firmware, is handled correctly it can still cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
HU02028 |
All |
Suggested
|
An issue, with timer cancellation, in the Remote Copy component may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using Remote Copy |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
IT19561 |
All |
Suggested
|
An issue with register clearance in the FC driver code may cause a node warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
Systems using 16Gb HBAs |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Reliability Availability Serviceability |
IT25457 |
All |
Suggested
|
Attempting to remove a copy of a volume, which has at least one image mode copy and at least one thin/compressed copy, in a Data Reduction Pool will always fail with a CMMVC8971E error
(show details)
Symptom |
None |
Environment |
Systems using Data Reduction Pools |
Trigger |
Try to remove a copy of a volume, which has at least one image mode copy and at least one thin/compressed copy, in a Data Reduction Pool |
Workaround |
Use svctask splitvdiskcopy to create a separate volume from the copy that should be deleted. This leaves the original volume with a single copy and creates a new volume from the copy that was split off. Then remove the newly created volume. |
|
8.2.1.0 |
Data Reduction Pools |
IT25970 |
All |
Suggested
|
After a FlashCopy consistency group is started a node may warmstart
(show details)
Symptom |
Single Node Warmstart |
Environment |
All systems |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
FlashCopy |
IT26049 |
All |
Suggested
|
An issue with CPU scheduling may cause the GUI to respond slowly
(show details)
Symptom |
None |
Environment |
Systems running v7.8 or later |
Trigger |
None |
Workaround |
None |
|
8.2.1.0 |
Graphical User Interface |