Release Note for systems built with IBM Spectrum Virtualize


This is the release note for the 8.2.1 release and details the issues resolved in all Program Temporary Fixes (PTFs) between 8.2.1.0 and 8.2.1.18. This document will be updated with additional information whenever a PTF is released.

This document was last updated on 28 October 2024.

  1. New Features
  2. Known Issues and Restrictions
  3. Issues Resolved
    1. Security Issues Resolved
    2. APARs Resolved
  4. Useful Links
Note. Detailed build version numbers are included in the Update Matrices in the Useful Links section

1. New Features

The following new features have been introduced in the 8.2.1 release: The following new feature has been introduced in the 8.2.1.3 release:

2. Known Issues and Restrictions

Note: For clarity, the terms "node" and "canister" are used interchangeably.
Details Introduced

Customers running v8.2.1.0 or lower, that upgrade to v8.2.1.16 or higher, may be unable to access the system using SSH if using an older SSH client that does not support modern key exchange algorithms.

The systems SSH protocol level will default to 3 when the upgrade completes, which will restrict the use of less secure algorithms. Use the GUI to change the SSH protocol level to a lower value after the upgrade if you use an older SSH client.

8.2.1.16

Due to an issue in the automatic upgrade process, customers with 32GB DH8 systems, running v8.2.1.4 or v8.2.1.5, cannot upgrade to v8.2.1.6 or later.

There is a manual upgrade option for these systems. Affected customers, wishing to use this option, should contact IBM Support for assistance quoting APAR HU02032

This is a restriction that may be lifted in a future PTF.

8.2.1.5

Customers with more than 5 x non-NVMe over FC hosts (i.e FC SCSI or iSCSI) in an I/O group must not attach any NVMe over FC hosts to that I/O group.
Customers with more than 20 x non-NVMe over FC hosts (i.e FC SCSI or iSCSI) in a cluster must not attach any NVMe over FC hosts to that cluster.

For new clusters without any hosts please refer to the appropriate v8.2.1 Configuration Limits and Restrictions pages for details of the maximum number of hosts that can be attached.

These limits will not be policed by the Spectrum Virtualize software. Any configurations that exceed these limits will experience significant adverse performance impact.

These limits will be lifted in a future major release.

8.2.1.0

Customers using Transparent Cloud Tiering should not upgrade to v8.2.1.0.

This is a restriction that may be lifted in a future PTF.

8.2.1.0

Spectrum Virtualize for Public Cloud v8.2.1 is not available.

8.2.1.0

Customers using iSCSI to virtualize backend controllers should not upgrade to v8.2.0 or later

This is a restriction that may be lifted in a future PTF.

8.2.0.0

Customers upgrading systems with more than 64GB of RAM to v8.1 or later will need to run chnodehw to enable access to the extra memory above 64GB.

Under some circumstances it may also be necessary to remove and re-add each node in turn.

8.1.0.0

Validation in the Upload Support Package feature will reject the new case number format in the PMR field.

This is a known issue that may be lifted in a future PTF. The fix can be tracked using APAR HU02392.

7.8.1.0

Systems, with NPIV enabled, presenting storage to SUSE Linux Enterprise Server (SLES) or Red Hat Enterprise Linux (RHEL) hosts running the ibmvfc driver on IBM Power can experience path loss or read-only file system events.

This is cause by issues within the ibmvfc driver and VIOS code.

Refer to this troubleshooting page for more information.

n/a
If an update stalls or fails then contact IBM Support for further assistance n/a
The following restrictions were valid but have now been lifted

Customers with direct attached external storage controllers cannot upgrade to v8.2.1.6.

This has been resolved, under APAR HU02077, in v8.2.1.8.

Please note that v8.2.1.5, or earlier, is not exposed to this restriction.

8.2.1.6

Systems containing FlashCore Modules (FCMs), running the v1.1.0 firmware level, are currently unable to perform software updates.

If the system is currently running 8.2.1.4, or later, then please upgrade the FCM firmware, to v1.2.7, before upgrading the system firmware.

If the system is running 8.2.1.3 or earlier - the restriction was temporary and has been lifted.

8.2.1.6

With Gemalto SafeNet KeySecure, the chkeyserverkeysecure -username <username> command is used to set the KeySecure username credential. If this is changed to a username that is not recognised by the key server to be the valid username, associated with the Spectrum Virtualize encryption key, then a subsequent re-key operation can cause key servers to appear offline.

This issue has been resolved in PTF v8.2.1.1.

8.2.1.0

A rare issue exists in the V5100, V7000 Gen3 and FlashSystem 9100 that can cause drives to become degraded during node warmstarts and upgrades. If this happens during the first half of a software upgrade - this can lead to a loss of access to data during the second half of the upgrade.

An enhancement was made in 8.2.1.8 and 8.3.0.0 to prevent the loss of access to data during upgrades, but this enhancement is not active during upgrades to these levels, only upgrades from them.

Clients on the exposed hardware and software configurations may wish to consider performing an upgrade that automatically pauses at the mid way to avoid the loss of access. More details can be found in the Spectrum Virtualize upgrade mid way point pausing How To.

This issue has been resolved in PTF v8.2.1.11.

8.2.0.0

3. Issues Resolved

This release contains all of the fixes included in the 8.1.3.1 release, plus the following additional fixes.

A release may contain fixes for security issues, fixes for APARs or both. Consult both tables below to understand the complete set of fixes included in the release.

3.1 Security Issues Resolved

Security issues are documented using a reference number provided by "Common Vulnerabilities and Exposures" (CVE).
CVE Identifier Link for additional Information Resolved in
CVE-2023-21930 7065011 8.2.1.18
CVE-2023-21937 7065011 8.2.1.18
CVE-2023-21938 7065011 8.2.1.18
CVE-2022-21626 6858041 8.2.1.17
CVE-2022-43873 6858047 8.2.1.17
CVE-2022-0778 6622017 8.2.1.16
CVE-2021-35603 6622019 8.2.1.16
CVE-2021-35550 6622019 8.2.1.16
CVE-2018-25032 6622021 8.2.1.16
CVE-2021-38969 6584337 8.2.1.15
CVE-2021-29873 6497111 8.2.1.14
CVE-2020-2781 6445063 8.2.1.12
CVE-2020-13935 6445063 8.2.1.12
CVE-2020-14577 6445063 8.2.1.12
CVE-2020-14578 6445063 8.2.1.12
CVE-2020-14579 6445063 8.2.1.12
CVE-2019-5544 6250889 8.2.1.11
CVE-2019-2964 6250887 8.2.1.11
CVE-2019-2989 6250887 8.2.1.11
CVE-2018-12404 6250885 8.2.1.11
CVE-2019-11477 1164286 8.2.1.6
CVE-2019-11478 1164286 8.2.1.6
CVE-2019-11479 1164286 8.2.1.6
CVE-2019-2602 1073958 8.2.1.6
CVE-2018-3180 ibm10884526 8.2.1.4
CVE-2018-12547 ibm10884526 8.2.1.4
CVE-2008-5161 ibm10874368 8.2.1.2
CVE-2018-5391 ibm10872368 8.2.1.2
CVE-2018-11776 ibm10741137 8.2.1.0
CVE-2017-17833 ibm10872546 8.2.1.0
CVE-2018-11784 ibm10872550 8.2.1.0
CVE-2018-5732 ibm10741135 8.2.1.0
CVE-2018-1517 ibm10872456 8.2.1.0
CVE-2018-2783 ibm10872456 8.2.1.0
CVE-2018-12539 ibm10872456 8.2.1.0
CVE-2018-1775 ibm10872486 8.2.1.0

3.2 APARs Resolved

Show details for all APARs
APAR Affected Products Severity Description Resolved in Feature Tags
HU02327 All HIPER Using addvdiskcopy in conjunction with expandvdisk with format may result in the original being overwritten, by the new copy, producing blank copies. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems running v8.2.1 or later
Trigger Using addvdiskcopy and expandvdisk with format
Workaround Wait until the format is completed before adding a copy
8.2.1.15 Volume Mirroring
HU02400 All HIPER A problem in the virtualization component of the system can cause a migration IO to be submitted in an incorrect context resulting in a node warmstart. In some cases it is possible that this IO has been submitted to an incorrect location on the backend, which can cause data corruption of an isolated small area (show details)
Symptom Data Integrity Loss
Environment Systems running v8.2.1 or later
Trigger None
Workaround None
8.2.1.15 Storage Virtualisation
HU02342 All Critical Occasionally when an offline drive returns to online state later than its peers in the same RAID array there can be multiple node warmstarts that send nodes into a service state (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8 or later
Trigger None
Workaround None
8.2.1.15 RAID
HU02374 SVC, V5000, V7000 Critical Hosts with Emulex 16Gbps HBAs may become unable to communicate with a system with 8Gbps Fibre Channel ports, after the host HBA is upgraded to firmware version 12.8.364.11. This does not apply to systems with 16Gb or 32Gb Fibre Channel ports (show details)
Symptom Loss of Access to Data
Environment Systems with 8Gbps Fibre Channel ports
Trigger Host Emulex 16Gbps HBA is upgraded to firmware version 12.8.364.11
Workaround Do not upgrade host HBA to firmware version 12.8.364.11
8.2.1.15 Hosts
HU02393 All Critical Automatic resize of compressed/thin volumes may fail causing warmstarts on both nodes in an I/O group (show details)
Symptom Loss of Access to Data
Environment Systems running v8.1 or later
Trigger None
Workaround None
8.2.1.15 Storage Virtualisation
HU02401 All Critical EasyTier can move extents between identical mdisks until one runs out of space (show details)
Symptom Loss of Access to Data
Environment Systems running v8.1 or later using EasyTier
Trigger None
Workaround Disable EasyTier. Manually migrate extents between mdisks
8.2.1.15 EasyTier
HU02406 All Critical An interoperability issue between Cisco NX-OS firmware and the Spectrum Virtualize Fibre Channel driver can cause a node warmstart on NPIV failback (for example during an upgrade) with the potential for a loss of access. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems using NPIV that are connected to Cisco SAN equipment running NX-OS 8.4(2c) or later
Trigger Initiate an NPIV failback operation by, for example, performing an upgrade
Workaround Disable NPIV (which will require any hot spare nodes to be removed first)
8.2.1.15 Interoperability
IT38015 All High Importance During RAID rebuild or copyback on systems with 16gb or less of memory, cache handling can lead to a deadlock which results in timeouts (show details)
Symptom Performance
Environment Systems with 16GB or less of memory
Trigger None
Workaround None
8.2.1.15 RAID
HU02366 All Suggested Slow internal resource reclamation by the RAID component can cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v8.2.1 or later
Trigger None
Workaround None
8.2.1.15 RAID
HU02433 FS5000, SVC, V5000, V5100, V7000 Suggested When a BIOS upgrade occurs excessive tracefile entries can be generated (show details)
Symptom None
Environment Gen 1 & 2 systems
Trigger None
Workaround None
8.2.1.15 System Update
HU02186 FS9100, V5100, V7000 HIPER NVMe drive pulls or firmware upgrades may lead to offline pools with the possibility of a small loss of data integrity. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems running v8.2.1, or later with NVMe drives
Trigger None
Workaround None
8.2.1.13 RAID
HU02186 (reverted) FS9100, V5100, V7000 HIPER This APAR has been reverted at this PTF. This APAR will be re-applied in a future PTF 8.2.1.12 RAID
HU02277 All HIPER RAID parity scrubbing can become stalled causing an accumulation of media errors leading to multiple drive failures with the possibility of data integrity loss. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems with model MZILS3T8HMLH read intensive SSDs at drive firmware MS24 are particularly susceptible to the data integrity (DI) issue. Other drive types may see multiple failures without DI issue
Trigger None
Workaround None
8.2.1.12 RAID
HU02313 FS9100, V5100, V7000 HIPER When a FlashCore Module (FCM) fails there is a chance that this can trigger other FCMs in the same control enclosure to also fail. If enough additional drives fail, at the same time, this can take the array offline and cause a loss of access to data. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2 or later using Flash Core Modules
Trigger None
Workaround None
8.2.1.12 Drives
HU01968 & HU02215 All Critical An upgrade may fail due to corrupt hardened data in a node. This can affect an I/O group (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround None
8.2.1.12 System Update
HU02058 All Critical Changing a remote copy relationship from GMCV to MM or GM can result in a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems using remote copy
Trigger None
Workaround None
8.2.1.12 Global Mirror, Global Mirror With Change Volumes, Metro Mirror
HU02184 All Critical When a 3PAR controller experiences a fault that prevents normal I/O processing it may issue a SCSI TARGET RESET command. This command is not supported and may cause multiple node asserts, possibly cluster-wide (show details)
Symptom Loss of Access to Data
Environment Systems with 3PAR backend controllers
Trigger None
Workaround None
8.2.1.12 Backend Storage
HU02213 SVC Critical A Hot Spare Node (HSN) timing window issue can, during an HSN activation or deactivation, cause the cluster to broadcast an invalid VPD update to other clusters on the SAN. This may trigger a Tier 2 recovery on the other cluster. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment SVC systems, with Hot Spare Nodes, using remote copy partnerships
Trigger None
Workaround Prior to an upgrade, or node hardware maintenance, remove the HSN
8.2.1.12 Hot Spare Node
HU02266 All Critical An issue in auto-expand can cause expansion to fail and the volume to be taken offline (show details)
Symptom Offline Volumes
Environment Systems running v8.2.1 or later using thin-provisioning
Trigger None
Workaround None
8.2.1.12 Thin Provisioning
HU02295 SVC Critical When upgrading from v8.2.1 or v8.3, in the presence of hot spare nodes, an issue with the handling of node metadata may cause a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment SVC systems running v8.2.1 or v8.3 with Hot Spare Node
Trigger None
Workaround None
8.2.1.12 System Update
HU02429 All Critical System can go offline shortly after changing the SMTP settings using the chemailserver command via the GUI (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8.1 or later
Trigger None
Workaround None
8.2.1.12 System Monitoring
HU02132 All High Importance Removing a thin-provisioned volume and then immediately creating one of the same size may cause node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems using thin-provisioned volumes
Trigger None
Workaround None
8.2.1.12 Thin Provisioning
HU02156 All High Importance Global Mirror environments may experience more frequent 1920 events due to writedone message queuing (show details)
Symptom Performance
Environment Systems using Global Mirror
Trigger None
Workaround None
8.2.1.12 Global Mirror
HU02164 All High Importance An issue in Remote Copy may cause a loss of hardened data when a node is warmstarted (show details)
Symptom Loss of Redundancy
Environment Systems using remote copy
Trigger None
Workaround None
8.2.1.12 Global Mirror, Global Mirror With Change Volumes, Metro Mirror
HU02176 All High Importance During upgrade a node may limit the number of target ports it reports causing a failover contradiction on hosts (show details)
Symptom Loss of Redundancy
Environment Systems with 12 or more Fibre Channel ports presenting storage to host type "tpgs"
Trigger None
Workaround None
8.2.1.12 Hosts
HU02200 All High Importance When upgrading from v8.1 or earlier to v8.2.1 or later a remote copy issue may cause a node warmstart, stalling the upgrade (show details)
Symptom Single Node Warmstart
Environment Systems running v8.1 or earlier using remote copy
Trigger None
Workaround None
8.2.1.12 System Update
HU02201 & HU02221 All High Importance Shortly after upgrading drive firmware, specific drive models can fail due to Too many long IOs to drive for too long errors (show details)
Symptom Loss of Redundancy
Environment Systems with the following drive models:
  • ST300MM0009 (300GB) - B5B8
  • ST600MM0009 (600GB) - B5B8
  • ST900MM0009 (900GB) - B5B8
  • ST1200MM0009 (1200GB) - B5B8
  • ST1200MM0129 (1800GB) - B5C9
  • ST2400MM0129 (2400GB) - B5C9
  • ST300MP0006 (300GB) - B6AA
  • ST600MP0006 (600GB) - B6AA
  • ST900MP0146 (900GB) - B6CB
Trigger None
Workaround None
8.2.1.12 Drives
HU02227 FS9100, SVC, V5100, V7000 High Importance Certain I/O patterns can cause compression hardware to post errors. When those errors exceed a threshold the node can be taken offline (show details)
Symptom Loss of Redundancy
Environment Systems running v8.2 or later using compressed volumes
Trigger None
Workaround None
8.2.1.12 Compression
HU02288 All High Importance A node might fail to come online after a reboot or warmstart such as during an upgrade (show details)
Symptom Loss of Redundancy
Environment Systems in a Stretched or HyperSwap topology
Trigger None
Workaround None
8.2.1.12 Reliability Availability Serviceability
HU02048 All Suggested An issue in the handling of ATS commands from VMware hosts can cause a single node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8, or later, presenting volumes to VMware hosts
Trigger None
Workaround None
8.2.1.12 Hosts
HU02095 All Suggested The effective_used_capacity field of lsarray/lsmdisk commands should be empty for RAID arrays which do not contain overprovisioned drives. However, sometimes this field can be zero even though it should be empty. This can cause incorrect provisioned capacity reporting in the GUI (show details)
Symptom None
Environment Systems with non-FCM arrays
Trigger None
Workaround None
8.2.1.12 Graphical User Interface
HU02142 All Suggested It is possible for a backend unmap process to become stalled, preventing system configuration changes from completing (show details)
Symptom Configuration
Environment Systems running v8.1.0, or later, using DRAID
Trigger None
Workaround None
8.2.1.12 Distributed RAID
HU02157 All Suggested Issuing a mkdistributedarray command may result in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using DRAID
Trigger None
Workaround None
8.2.1.12 Distributed RAID
HU02241 All Suggested IP Replication can fail to create IP partnerships via the secondary cluster management IP (show details)
Symptom None
Environment All systems
Trigger None
Workaround Use primary management IP to run mkippartnership commands
8.2.1.12 IP Replication
HU02244 SVC Suggested False positive node error 766 (depleted CMOS battery) messages may appear in the Event Log (show details)
Symptom None
Environment SVC systems with SV1 model nodes running v8.2.1 or later
Trigger None
Workaround None
8.2.1.12 System Monitoring
HU02292 & HU02308 All Suggested The use of maximum replication delay within Global Mirror may occasionally cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Remote Copy
Trigger None
Workaround None
8.2.1.12 Global Mirror
HU02332 & HU02336 All Suggested When an I/O is received, from a host, with invalid or inconsistent SCSI data but a good checksum it may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8 or later
Trigger None
Workaround None
8.2.1.12 Hosts
HU02354 SVC Suggested An issue in the handling of read transfers may cause hung host IOs leading to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v8.2.1 or later
Trigger None
Workaround None
8.2.1.12 Hosts
HU02358 All Suggested An issue in Remote Copy, that stalls a switch of direction, can cause I/O timeouts leading to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Remote Copy
Trigger None
Workaround None
8.2.1.12 Global Mirror, Global Mirror With Change Volumes, Metro Mirror
HU01894 All HIPER After node reboot, or warmstart, some volumes accessed by AIX, VIO or VMware hosts may experience stuck SCSI2 reservations on the NPIV failover ports of the partner node. This can cause a loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems using NPIV to present storage to AIX, VIO or VMware hosts
Trigger None
Workaround Clear reservation by either:
  • Unmap & re-map volume;
  • LUN reset from host.
8.2.1.11 Hosts
HU02141 All HIPER An issue in the max replication delay function may trigger a Tier 2 recovery, after posting multiple 1920 errors in the Event Log. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems using remote copy
Trigger None
Workaround Set the max_replication_delay value to 0 (disabled)
8.2.1.11 Global Mirror
HU02186 (reverted in 8.2.1.12) FS9100, V5100, V7000 HIPER NVMe drive pulls or firmware upgrades may lead to offline pools with the possibility of a small loss of data integrity. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems running v8.2.1, or later with NVMe drives
Trigger None
Workaround None
8.2.1.11 RAID
HU02205 All HIPER Incremental FlashCopy targets can be corrupted when the FlashCopy source is a target of a remote copy relationship (show details)
Symptom Data Integrity Loss
Environment Systems using Incremental FlashCopy with remote copy
Trigger None
Workaround None
8.2.1.11 FlashCopy, Global Mirror, Global Mirror With Change Volumes, Metro Mirror
HU02212 All HIPER Remote Copy secondary may have inconsistent data following a stop with -access due to a missing bitmap merge from FlashCopy to Remote Copy. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems running v8.2.1 or later using GMCV or HyperSwap
Trigger None
Workaround None
8.2.1.11 Global Mirror With Change Volumes, HyperSwap
HU02237 All HIPER Under a rare and complicated set of conditions, a RAID 1 or RAID 10 array may drop a write, causing undetected data corruption. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems using RAID 1 or RAID 10 arrays
Trigger None
Workaround None
8.2.1.11 RAID
HU02238 All HIPER Force-stopping a FlashCopy map, where the source volume is a Metro or Global Mirror target volume, may cause other FlashCopy maps to return invalid data if they are not 100% copied, in specific configurations. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems using FlashCopy
Trigger None
Workaround None
8.2.1.11 FlashCopy, Global Mirror, Metro Mirror
HU01921 All Critical Where FlashCopy mapping targets are also in remote copy relationships there may be node warmstarts with a temporary loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems using FlashCopy with remote copy
Trigger None
Workaround If one reverse FlashCopy mapping has been stopped and another FlashCopy mapping, to the same target, is to be started, then delete the first reverse FlashCopy mapping before starting the second
8.2.1.11 FlashCopy, Global Mirror, Metro Mirror
HU01924 All Critical Migrating extents to an MDisk, that is not a member of an MDisk group, may result in a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Migrate extents to an MDisk, that is not a member of an MDisk group
Workaround Only specify a target MDisk that is part of the same MDisk group as the volume copy having extents migrated
8.2.1.11 Thin Provisioning
HU01970 All Critical When a GMCV relationship is stopped, with the -access option, and the secondary volume is immediately deleted with -force, then all nodes may repeatedly warmstart (show details)
Symptom Loss of Access to Data
Environment Systems using GMCV
Trigger Stop a GMCV relationship with -access and immediately delete the secondary volume
Workaround Do not remove secondary volume, with -force, if the backward FC map from the secondary change volume to the secondary volume is still in progress
8.2.1.11 Global Mirror With Change Volumes
HU02005 All Critical An issue in the background copy process prevents grains, above a 128TB limit, from being cleaned properly. As a consequence there may be multiple node warmstarts with the potential for a loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems using remote copy
Trigger Volumes greater than 128TB being added to remote copy relationships
Workaround The maximum size of volumes in remote copy relationships should be limited to 128TB
8.2.1.11 Global Mirror, Global Mirror With Change Volumes, Metro Mirror
HU02054 All Critical The event log handler maintains a second list of events. On rare occasions, for log full events, these lists can get out of step, resulting in a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8 or later
Trigger None
Workaround If an error event log full message (1002) is presented clear the event log, rather than marking that event as fixed
8.2.1.11 System Monitoring
HU02065 All Critical Mishandling of Data Reduction Pool allocation request rejections can lead to node warmstarts that can take an MDisk group offline (show details)
Symptom Loss of Access to Data
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.11 Data Reduction Pools
HU02069 All Critical When a SCSI command, containing an invalid byte, is received there may be a node warmstart. This can affect both nodes, in an I/O group, at the same time (show details)
Symptom Loss of Access to Data
Environment Systems with hosts using Fibre Channel connectivity
Trigger None
Workaround None
8.2.1.11 Hosts
HU02097 All Critical Workloads, with data that is highly suited to deduplication, can provoke high CPU utilisation, as multiple destinations try to dedupe to one source. This adversely impacts performance with the possibility of offline MDisk groups (show details)
Symptom Loss of Access to Data
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.11 Data Reduction Pools
HU02106 All Critical Multiple node warmstarts, in quick succession, can cause the partner node to lease expire (show details)
Symptom Loss of Access to Data
Environment Systems using IP Quorum or NVMe drives as quorum devices
Trigger None
Workaround None
8.2.1.11 IP Quorum, Quorum
HU02108 All Critical Deleting a managed disk group, with -force, may cause multiple warmstarts with the possibility of a loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.11 Data Reduction Pools
HU02109 All Critical Free extents may not be unmapped after volume deletion, or migration, resulting in out-of-space conditions on backend controllers (show details)
Symptom Loss of Access to Data
Environment Systems running v8.3.0 or later
Trigger None
Workaround None
8.2.1.11 Backend Storage, SCSI Unmap
HU02135 All Critical Removing multiple IQNs for an iSCSI host can result in a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2 or later with iSCSI connected hosts
Trigger Use a single rmhostport command to remove multiple IQN from an iSCSI host
Workaround Remove iSCSI host ports one IQH at a time
8.2.1.11 iSCSI
HU02138 All Critical An issue in Data Reduction Pool garbage collection can cause I/O timeouts leading to an offline pool (show details)
Symptom Offline Volumes
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.11 Data Reduction Pools
HU02154 All Critical If a node is rebooted, when remote support is enabled, then all other nodes will warmstart (show details)
Symptom Loss of Access to Data
Environment Systems running v8.1 or later using remote support
Trigger With remote support enabled, reboot a node using the 'satask stopnode -reboot <node id>' command
Workaround Temporarily disable remote support when rebooting a node using 'chsra -remotesupport disable'
8.2.1.11 Support Remote Assist
HU02155 All Critical Upgrading to v8.2.1 may result in offline managed disk groups and OOS events (1685/1687) appearing in the Event Log (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2.0, or earlier, using Data Reduction Pools
Trigger Upgrading to v8.2.1
Workaround None
8.2.1.11 Data Reduction Pools
HU02197 All Critical Bulk volume removals can adversely impact related FlashCopy mappings leading to a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems running v7.7.1, or later, using FlashCopy
Trigger None
Workaround None
8.2.1.11 FlashCopy
HU02222 All Critical Where the source volume of an incremental FlashCopy map is also a Metro or Global Mirror target volume that is using a change volume or is a Hyperswap volume, then there is a possibility that not all data will be copied to the FlashCopy target. For more details refer to this Flash (show details)
Symptom Data Integrity Loss
Environment Systems using Remote Copy
Trigger None
Workaround None
8.2.1.11 Global Mirror With Change Volumes
IT25367 All Critical A T2 recovery may occur when an attempt is made to upgrade, or downgrade, the firmware for an unsupported drive type (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Attempt to upgrade/downgrade the firmware for an unsupported drive type
Workaround None
8.2.1.11 Drives
IT31113 All Critical After a manual power off and on, of a system, both nodes, in an I/O group, may repeatedly assert into a service state (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2 or later
Trigger Manual power off and on of a system whilst a RAID rebuild is in progress
Workaround None
8.2.1.11 RAID
HU01923 All High Importance An issue in the way Global Mirror handles write sequence numbers >512 may cause multiple node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems using GM
Trigger None
Workaround None
8.2.1.11 Global Mirror
HU02080 All High Importance When a Data Reduction Pool is running low on free space, the credit allocation algorithm, for garbage collection, can be exposed to a race condition, adversely affecting performance (show details)
Symptom Performance
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.11 Data Reduction Pools
HU02114 FS5000, FS9100, V7000 High Importance Upgrading FCM firmware on multiple I/O group systems can cause a drive to become stuck at 0% sync with the corresponding array in a 'syncing' state (show details)
Symptom Performance
Environment Multiple I/O group systems with Flash Core Modules
Trigger None
Workaround None
8.2.1.11 Drives
HU02123 All High Importance For direct-attached hosts, a race condition between the FLOGI and Link UP processes can result in FC ports not coming online (show details)
Symptom Loss of Redundancy
Environment Systems with direct-attached hosts
Trigger None
Workaround None
8.2.1.11 Hosts
HU02149 SVC High Importance When an Enhanced Stretch Cluster is using NPIV, in transitional mode, the path priority is not being reported correctly to some hosts (show details)
Symptom Performance
Environment Systems in an Enhanced Stretch Cluster topology that are using NPIV in its transitional mode
Trigger None
Workaround Manually set the preferred path if possible within the host's MPIO settings. Run NPIV in enabled or disabled mode
8.2.1.11 Hosts
HU02168 V5000, V7000 High Importance In the event of unexpected power loss a node may not save system data (show details)
Symptom Loss of Redundancy
Environment Storwize V5000 Gen2, V7000 Gen 2 and Gen 2+ systems
Trigger Sudden power loss
Workaround When shutting down always use the CLI, service GUI or management GUI. Do not use removal of electrical supply
8.2.1.11 Reliability Availability Serviceability
HU02203 FS9100, V5000, V7000 High Importance When a node reboots, it is possible for the node to be unable to communicate with some of the NVMe drives in the enclosure (show details)
Symptom Loss of Redundancy
Environment Systems with NVMe drives
Trigger Drive reseat or node reboot
Workaround None
8.2.1.11 Drives
HU01868 All Suggested After deleting an encrypted external MDisk, it is possible for the encrypted status of volumes to change to no, even though all remaining MDisks are encrypted (show details)
Symptom None
Environment Systems using encryption
Trigger Delete an encrypted external MDisk
Workaround Ensure that all MDisks in the MDisk group are encrypted - this will ensure that data is encrypted
8.2.1.11 Encryption
HU01917 All Suggested Chrome browser support requires a self-signed certificate to include subject alternate name (show details)
Symptom None
Environment Systems accessed using the Chrome browser
Trigger None
Workaround Accept invalid certificate
8.2.1.11 Graphical User Interface
HU01931 SVC, V7000 Suggested Where a high rate of CLI commands are received, it is possible for inter-node processing code to be delayed which results in a small increase in receive queue time on the config node (show details)
Symptom Performance
Environment SVC and Storwize V7000 systems
Trigger None
Workaround If CPU utilisation is less than 40% then creating a compressed volume may reduce response times
8.2.1.11 Performance
HU02015 FS9100, V5000, V7000 Suggested Some read-intensive SSDs are incorrectly reporting wear rate thresholds generating unnecessary errors in the Event Log (show details)
Symptom None
Environment Systems using Toshiba M4 Read-Intensive SSDs
Trigger None
Workaround None
8.2.1.11 Drives
HU02091 V5000 Suggested Upgrading to v8.2.1.8, or later, may result in a licensing error in the Event Log (show details)
Symptom None
Environment Lenovo Storage V Series systems
Trigger Upgrade to v8.2.1.8 or later
Workaround None
8.2.1.11 Licensing
HU02103 FS9100, V5000, V7000 Suggested The system management firmware may, incorrectly, attempt to obtain an IP address, using DHCP, making it accessible via Ethernet (show details)
Symptom None
Environment FlashSystem 9100, Storwize V7000 Gen 3 and Storwize V5100 systems
Trigger None
Workaround None
8.2.1.11
HU02111 All Suggested An issue with how Data Reduction Pool handles data, at the sub-extent level, may result in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.11 Data Reduction Pools
HU02124 All Suggested Due to an issue with FCM thin provisioning calculations the GUI may incorrectly display volume capacity and capacity savings as zero (show details)
Symptom None
Environment Systems running v8.2.1 or later
Trigger None
Workaround None
8.2.1.11 System Monitoring
HU02137 All Suggested An issue with support for target resets in Nimble Storage controllers may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v8.2, or later, with Nimble Storage backend controllers
Trigger None
Workaround None
8.2.1.11 Backend Storage
HU02173 All Suggested During a pending fabric login, when an abort is received, it is possible for a related entry in the WWPN table to not be removed. The node will warmstart to clear this condition (show details)
Symptom Single Node Warmstart
Environment Systems running v8.2 or later
Trigger None
Workaround None
8.2.1.11 Reliability Availability Serviceability
HU02183 All Suggested An issue in the way inter-node communication is handled can lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v8.2 or later
Trigger None
Workaround None
8.2.1.11 Reliability Availability Serviceability
HU02190 All Suggested Error 1046 not triggering a Call Home even though it is a hardware fault (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
8.2.1.11 System Monitoring
HU02214 All Suggested Under a certain I/O pattern it is possible for metadata management in Data Reduction Pools to become inconsistent leading to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.11 Data Reduction Pools
HU02247 All Suggested Unnecessary Ethernet MAC flapping messages reported in switch logs (show details)
Symptom None
Environment FlashSystem 9100, Storwize V7000 Gen 3 and V5100 systems
Trigger None
Workaround None
8.2.1.11 Reliability Availability Serviceability
HU02143 All High Importance The performance profile, for some enterprise tier drives, may not correctly match the drives capabilities leading to that tier being overdriven (show details)
Symptom Performance
Environment Systems running v8.2 or later using EasyTier. Note: This issue does not affect DRAID 5 arrays with stripe width of 8 or 9, or DRAID6 arrays with stripe width of 10 or 12.
Trigger None
Workaround None
8.2.1.10 EasyTier
HU02104 All HIPER An issue in the RAID component, in the presence of very high I/O workload and the exhaustion of cache resources, can see a deadlock condition occurring which prevents further I/O processing. The system detects this issue and takes the storage pool offline for a six minute period, to clear the problem. The pool is then brought online automatically, and normal operation resumes. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2.1 or later
Trigger None
Workaround Consider using a pool throttle to limit the I/O throughput
8.2.1.9 RAID
HU02133 FS9100, V5000, V7000 High Importance NVMe drives may become degraded after a drive reseat or node reboot (show details)
Symptom None
Environment Systems with NVMe drives
Trigger Drive reseat or node reboot
Workaround None
8.2.1.9 Drives
HU02102 All Suggested Excessive processing time required for FlashCopy bitmap operations, associated with large (> 20TB) Global Mirror change volumes, may lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Global Mirror with Change Volumes where some GMCV volumes are >20TB
Trigger None
Workaround Limit GMCV volume capacity to 20TB or less
8.2.1.9 Global Mirror With Change Volumes
HU02126 SVC, V5000, V7000 Suggested There is a low probability that excessive SSH connections may trigger a single node warmstart on the configuration node (show details)
Symptom Single Node Warmstart
Environment Systems with Gen 1 and 2 hardware
Trigger More than one SSH connection attempt per second might occasionally cause the config node to warmstart
Workaround Reduce frequency of SSH connections
8.2.1.9 Command Line Interface
HU02131 All Suggested When changing DRAID configuration, for an array with an active workload, a deadlock condition can occur resulting in a single node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v8.2.1, or later, using DRAID
Trigger None
Workaround None
8.2.1.9 Distributed RAID
HU02064 SVC, V7000 HIPER An issue in the firmware for compression accelerator cards can cause offline compressed volumes. For more details refer to this Flash (show details)
Symptom Offline Volumes
Environment Systems running v8.2.1.x, or later, using hardware compression
Trigger None
Workaround None
8.2.1.8 Compression
HU02083 All HIPER During DRAID rebuilds, an issue in the handling of memory buffers can lead to multiple node warmstarts and a loss of access to data. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2.1.6 or v8.3.0.0 using DRAID. Probability is highest for systems with an exact multiple of 48 drives and a stripe width of 16
Trigger None
Workaround None
8.2.1.8 Distributed RAID
HU01967 All Critical When I/O, in remote copy relationships, experiences delays (1720 and/or 1920 errors are logged) an I/O group may warmstart (show details)
Symptom Loss of Access to Data
Environment Systems using remote copy
Trigger Performance issues affecting replication I/O
Workaround Use a max replication delay value of 30 seconds or greater
8.2.1.8 Global Mirror, Global Mirror With Change Volumes, Metro Mirror
HU02036 All Critical It is possible for commands, that alter pool-level extent reservations (i.e. migratevdisk or rmmdisk), to conflict with an ongoing EasyTier migration, resulting in a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2 or later with EasyTier enabled
Trigger None
Workaround Disable EasyTier on the source pool; Wait 10 minutes, so that any ongoing EasyTier requests complete; Issue the migratevdisk/rmmdisk command; Enable EasyTier on the source pool.
8.2.1.8 EasyTier
HU02044 All Critical Multiple DRAID arrays can, where one is performing a rebuild, be exposed to a RAID deadlock condition resulting in multiple node warmstarts and a loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems using Distributed RAID with Data Reduction Pools
Trigger None
Workaround None
8.2.1.8 Data Reduction Pools, Distributed RAID
HU02050 FS9100, V5000, V7000 Critical Compression hardware can have an issue processing certain types of data resulting in node reboots and marking the compression hardware as faulty even though it is serviceable (show details)
Symptom Loss of Access to Data
Environment FlashSystem 9100, Storwize V5100 and V7000 Gen 3 systems
Trigger None
Workaround None
8.2.1.8 Compression
HU02063 All Critical HyperSwap clusters with only two surviving nodes may experience warmstarts on both of those nodes where rcbuffersize is set to 512MB (show details)
Symptom Loss of Access to Data
Environment Systems using Hyperswap
Trigger None
Workaround Reduce rcbuffersize to a value less than 512
8.2.1.8 HyperSwap
HU02077 All Critical A node upgrading to v8.2.1 or later will lose access to controllers directly-attached to its FC ports and the upgrade will stall (show details)
Symptom Loss of Access to Data
Environment Systems that are FC direct-attached to backend storage controllers
Trigger System upgrade
Workaround None
8.2.1.8 Backend Storage
HU02086 All Critical An issue, in IP Quorum, may cause a Tier 2 recovery, during initial connection to a candidate device (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2.1.0 or later that are using IP Quorum
Trigger None
Workaround None
8.2.1.8 IP Quorum
HU02089 All Critical Due to changes to quorum management, during an upgrade to v8.2.x, or later, there may be multiple warmstarts, with the possibility of a loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems running v8.1.3 or earlier with normal configurations of more than 6 nodes, or multi-site configurations of more than 4 nodes, and no external shared MDisks
Trigger Upgrading to v8.2.x or later
Workaround None
8.2.1.8 System Update
HU02121 All Critical When the system changes from copyback to rebuild a failure to clear related metadata can cause multiple node warmstarts, with the possibility of a loss of access (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2.1.x, or later, that are using DRAID
Trigger None
Workaround None
8.2.1.8 Distributed RAID
IT26257 All Critical Starting a relationship, when the remote volume is offline, may result in a T2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems using Hyperswap
Trigger None
Workaround None
8.2.1.8 HyperSwap
IT30595 All Critical A resource shortage in the RAID component can cause MDisks to be taken offline (show details)
Symptom Offline Volumes
Environment Systems running v8.1 or later
Trigger None
Workaround None
8.2.1.8 RAID
HU01836 All High Importance When an auxiliary volume is moved an issue with pausing the master volume can lead to node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems using Hyperswap
Trigger None
Workaround None
8.2.1.8 HyperSwap
HU01942 FS9100, V5000, V7000 High Importance NVMe drive ports can go offline, for a very short time, when an upgrade of that drives firmware commences (show details)
Symptom None
Environment FlashSystem 9100, Storwize V7000 Gen 3 and Storwize V5100 systems
Trigger Start a NVMe drive firmware upgrade
Workaround None
8.2.1.8 Drives
HU02049 All High Importance GUI session handling has an issue that can generate many exceptions, adversely impacting GUI performance (show details)
Symptom Performance
Environment All systems
Trigger None
Workaround None
8.2.1.8 Graphical User Interface
HU02078 SVC High Importance Heavily unbalanced workloads, in stretched-cluster configurations, can bias inter-node traffic through one port, adversely affecting performance (show details)
Symptom Performance
Environment SVC systems in a stretched-cluster configuration
Trigger None
Workaround Throttle or modify workloads if possible
8.2.1.8 Performance
HU01880 All Suggested When a write, to a secondary volume, becomes stalled, a node at the primary site may warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using remote copy
Trigger None
Workaround None
8.2.1.8 Global Mirror, Global Mirror With Change Volumes, Metro Mirror
HU01936 All Suggested When shrinking a volume, that has host mappings, there may be recurring node warmstarts (show details)
Symptom Single Node Warmstart
Environment Systems running v8.1 or later
Trigger Shrink a volume while it is mapped to a host
Workaround Remove all host mappings, for a volume, before performing shrinkvdisksize
8.2.1.8 Cache
HU02021 All Suggested Disabling garbage collection may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.8 Data Reduction Pools
HU02085 All Suggested Freeze time of Global Mirror remote copy consistency groups may not be updated correctly in certain scenarios (show details)
Symptom Single Node Warmstart
Environment Systems running v7.8 or later using Global Mirror
Trigger None
Workaround None
8.2.1.8 Global Mirror
HU02090 FS9100, V5000, V7000 Suggested When a failing drive experiences an error, RAID may mishandle it, resulting in a node warmstart (show details)
Symptom Single Node Warmstart
Environment FlashSystem 9100 and Storwize systems
Trigger None
Workaround None
8.2.1.8 RAID
HU02093 V5000 Suggested A locking issue in the inter-node communications, of V5030 systems, can lead to a deadlock condition, resulting in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Storwize V5030 systems running v8.2.1.0 or later
Trigger None
Workaround None
8.2.1.8 Reliability Availability Serviceability
HU02099 All Suggested Cloud callhome error 3201 messages may appear in the Event Log (show details)
Symptom None
Environment Systems running v8.2.1 or later
Trigger None
Workaround None
8.2.1.8 System Monitoring
IT30448 All Suggested If an IP Quorum app is killed, during the commit phase of a code upgrade, then that offline IP Quorum device cannot be removed, post upgrade (show details)
Symptom Configuration
Environment Systems running v8.1.3, or earlier, using IP Quorum
Trigger Upgrade to v8.2.0 or later, kill an IP Quorum app during the commit phase
Workaround None
8.2.1.8 IP Quorum
IT30449 V5000 Suggested Attempting to activate USB encryption on a new V5030E will fail with a CMMVCU6054E error (show details)
Symptom Configuration
Environment Storwize V5030E systems
Trigger Attempt activation of USB encryption
Workaround None
8.2.1.8 Encryption
HU01998 All HIPER All SCSI command types can set volumes as busy resulting in I/O timeouts and multiple node warmstarts, with the possibility of an offline I/O group. For more details refer to this Flash (show details)
Symptom Multiple Node Warmstarts
Environment All systems
Trigger None
Workaround None
8.2.1.6 Hosts
HU02014 SVC HIPER After a loss of power, where a node has a dead CMOS battery, it will fail to restart correctly. It is possible for both nodes in an I/O group to experience this issue (show details)
Symptom Loss of Access to Data
Environment SVC systems using SV1 model nodes
Trigger None
Workaround None
8.2.1.6 Reliability Availability Serviceability
HU01888 & HU01997 All Critical An issue with restore mappings, in the FlashCopy component, can cause an I/O group to warmstart (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround None
8.2.1.6 FlashCopy
HU01933 All Critical Under rare circumstances the Data Reduction Pool deduplication rehoming process can become truncated. Subsequent detection of inconsistent metadata can lead to offline Data Reduction Pools (show details)
Symptom Loss of Access to Data
Environment Systems running v8.1.3 or later using Deduplication
Trigger None
Workaround None
8.2.1.6 Data Reduction Pools, Deduplication
HU01985 All Critical As a consequence of a Data Reduction Pool recovery, bad metadata may be created. When the region of disk associated with the bad metadata is accessed there may be an I/O group warmstarts (show details)
Symptom Loss of Access to Data
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.6 Data Reduction Pools
HU01989 All Critical For large drives, bitmap scanning, during an array rebuild, can timeout resulting in multiple node warmstarts, possibly leading to offline I/O groups (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID with drives of 8TB or more
Trigger None
Workaround None
8.2.1.6 Distributed RAID
HU02016 SVC Critical A memory leak in the component that handles thin-provisioned MDisks can lead to an adverse performance impact with the possibility of offline MDisks. For more details refer to this Flash (show details)
Symptom Offline Volumes
Environment SVC systems
Trigger None
Workaround None
8.2.1.6 Backend Storage
HU02027 All Critical Fabric congestion can cause internal resource constraints, in 16Gb HBAs, leading to lease expiries (show details)
Symptom Loss of Access to Data
Environment Systems using 16Gb HBAs
Trigger Fabric congestion affecting local node-node traffic
Workaround Prevent fabric congestion that might affect local node-node connectivity
8.2.1.6 Reliability Availability Serviceability
HU02043 All Critical Collecting a snap can cause nodes to run out of boot drive space and go offline with node error 565 (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Taking many snap data collections on the same config node
Workaround Manually delete unneeded snaps from the boot drive
8.2.1.6 Support Data Collection
HU02045 All Critical When a node is removed from the cluster, using CLI, it may still be shown as online in the GUI. If an attempt is made to shutdown this node, from the GUI, whilst it appears to be online, then the whole cluster will shutdown (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Remove a node from the cluster using CLI. With the node showing as online, use the GUI to shut it down
Workaround Manually refresh GUI browser page, after removing a node via CLI
8.2.1.6 Graphical User Interface
HU01890 All High Importance FlashCopy mappings, from master volume to primary change volume, may become stalled when a T2 recovery occurs whilst the mappings are in a copying state (show details)
Symptom None
Environment Systems using Global Mirror with Change Volumes
Trigger None
Workaround None
8.2.1.6 Global Mirror With Change Volumes
HU02037 All High Importance A FlashCopy consistency group, with a mix of mappings in different states, cannot be stopped (show details)
Symptom None
Environment Systems using FlashCopy
Trigger Some, but not all, mappings, in a consistency group, have their target volumes run out of space
Workaround None
8.2.1.6 FlashCopy
HU02053 FS9100, V5100, V7000 High Importance An issue with canister BIOS update can stall system upgrades (show details)
Symptom Loss of Redundancy
Environment FS9100, V7000 Gen 3 and V5100 systems
Trigger Upgrade to v8.3.0
Workaround None
8.2.1.6 System Update
HU02055 All High Importance Creating a FlashCopy snapshot, in the GUI, does not set the same preferred node for both source and target volumes. This may adversely impact performance (show details)
Symptom Performance
Environment Systems using FlashCopy
Trigger None
Workaround Use the movevdisk command to manually set the same preferred node for both the source and target volumes in the FC map
8.2.1.6 FlashCopy
HU02072 All High Importance An issue in the handling of email transmission can write a large file to the node boot drive. If this causes the boot drive to become full, the node will go offline with error 565 (show details)
Symptom Loss of Redundancy
Environment Systems running v8.2.1 or later
Trigger None
Workaround None
8.2.1.6 System Monitoring
HU01843 All Suggested A node hardware issue can cause a CLI command to timeout resulting in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v8.1 or later
Trigger None
Workaround None
8.2.1.6 Command Line Interface
HU01892 All Suggested LUNs of greater than 2TB, presented by HP XP7 storage controllers, are not supported (show details)
Symptom Configuration
Environment Systems with HP XP7 backend controllers
Trigger None
Workaround None
8.2.1.6 Backend Storage
HU01974 All Suggested With all Remote Support Assistant connections closed, the GUI may show that a connection is still in progress (show details)
Symptom None
Environment Systems running v8.1 or later using Remote Support Assistance
Trigger None
Workaround None
8.2.1.6 System Monitoring
HU01978 All Suggested Unable to create HyperSwap volumes. The mkvolume command fails with CMMVC7050E error (show details)
Symptom None
Environment Systems running v8.2 or later using HyperSwap
Trigger None
Workaround Use the early (pre mkvolume) procedure for creating HyperSwap volumes
8.2.1.6 HyperSwap
HU01979 All Suggested The figure for used_virtualization, in the output of a lslicense command, may be unexpectedly large (show details)
Symptom None
Environment Systems running v8.1 or later
Trigger None
Workaround None
8.2.1.6 Command Line Interface
HU01982 All Suggested In an environment, with multiple IP Quorum servers, if the quorum component encounters a duplicate UID then a node may warmstart (show details)
Symptom Single Node Warmstart
Environment Systems running v8.2.1 or later running IP quorum
Trigger None
Workaround None
8.2.1.6 IP Quorum
HU01983 All Suggested Improve debug data capture to assist in determining the reason for a Data Reduction Pool to be taken offline (show details)
Symptom None
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.6 Data Reduction Pools
HU01986 All Suggested An accounting issue in the FlashCopy component may cause node warmstarts (show details)
Symptom Single Node Warmstart
Environment Systems using FlashCopy
Trigger None
Workaround None
8.2.1.6 FlashCopy
HU01991 All Suggested An issue in the handling of extent allocation, in the Data Reduction Pool component, can cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.6 Data Reduction Pools
HU02020 FS9100, V5000, V7000 Suggested An internal hardware bus, running at the incorrect speed, may give rise to spurious DIMM over-temperature errors (show details)
Symptom None
Environment FlashSystem 9100, Storwize V7000 Gen 3 and Storwize V5100 systems
Trigger None
Workaround None
8.2.1.6 Reliability Availability Serviceability
HU02029 All Suggested An issue with the SSMTP process may result in failed callhome, inventory reporting and user notifications. A testemail command will fail with a CMMVC9051E error (show details)
Symptom None
Environment Systems running v8.2.1 or later
Trigger None
Workaround None
8.2.1.6 System Monitoring
HU02039 All Suggested An issue in the management steps of Data Reduction Pool recovery may lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.6 Data Reduction Pools
HU02067 All Suggested If multiple recipients are specified, for callhome emails, then no callhome emails will be sent (show details)
Symptom None
Environment Systems running v8.2.1.5 or later
Trigger Specify multiple recipients for callhome email messages
Workaround None
8.2.1.6 System Monitoring
HU02129 All Suggested GUI drive filtering fails with An error occurred loading table data (show details)
Symptom None
Environment Systems running v8.2.1 or later
Trigger None
Workaround None
8.2.1.6 Graphical User Interface
HU02007 All HIPER During volume migration an issue, in the handling of old to new extents transfer, can lead to cluster-wide warmstarts (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround None
8.2.1.5 Storage Virtualisation
HU02009 All Critical Systems which are using Data Reduction Pools, with the maximum possible extent size of 8GB, and which experience a very specific I/O workload, may experience an issue due to garbage collection. This can cause repeated node warmstarts and loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.5 Data Reduction Pools
HU02011 All High Importance When a node warmstart occurs on a system using Data Reduction Pools, there is a small possibility that the node will not automatically return online. If the partner node is also offline, this can cause temporary loss of access to data (show details)
Symptom Loss of Redundancy
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.5 Data Reduction Pools
HU02012 All High Importance Under certain I/O workloads the garbage collection process can adversely impact volume write response times (show details)
Symptom Performance
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.5 Data Reduction Pools
HU01865 All HIPER When creating a HyperSwap relationship, using addvolumecopy (or similar methods), the system should perform a synchronisation operation to copy the data from the original copy to the new copy. In some rare cases this synchronisation is skipped, leaving the new copy with bad data (all zeros) (show details)
Symptom Data Integrity Loss
Environment Systems running v7.5 or later using HyperSwap
Trigger None
Workaround None
8.2.1.4 HyperSwap
HU01918 All HIPER Where Data Reduction Pools have been created on earlier code levels, upgrading the system, to an affected release, can cause an increase in the level of concurrent flushing to disk. This may result in a loss of access to data. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems running v8.1.3.4, v8.2.0.3 or v8.2.1.x using Data Reduction Pools
Trigger None
Workaround None
8.2.1.4 Data Reduction Pools
HU02008 All HIPER When a DRAID rebuild occurs, occasionally a RAID deadlock condition can be triggered by a particular type of I/O workload. This can lead to repeated node warmstarts and a loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems using Distributed RAID
Trigger None
Workaround None
8.2.1.4 Distributed RAID
HU01887 All Critical In circumstances where host configuration data becomes inconsistent, across nodes, an issue in the CLI policing code may cause multiple warmstarts (show details)
Symptom Loss of Access to Data
Environment Systems using Host Clusters
Trigger None
Workaround None
8.2.1.4 Command Line Interface, Host Cluster
HU01900 All Critical Executing a command, that can result in a shared mapping being created or destroyed, for an individual host, in a host cluster, without that command applying to all hosts in the host cluster, may lead to multiple node warmstarts with the possibility of a T2 recovery (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Execute a command that can result in a shared mapping being created or destroyed
Workaround Do not use the -force option when running a rmhostiogrp command
8.2.1.4 Host Cluster
HU01910 All Critical When FlashCopy mappings are created, with a grain size of 64KB, it is possible for an overflow condition in the bitmap to occur. This can resulting in multiple node warmstarts with a possible loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems using FlashCopy mappings with a 64KB grain size
Trigger None
Workaround Select a grain size of 256KB when creating FlashCopy mappings
8.2.1.4 FlashCopy
HU01928 All Critical When two IOs attempt to access the same address, the state of the data may be incorrectly set to invalid causing offline volumes and, possibly, offline pools (show details)
Symptom Offline Volumes
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.4 Data Reduction Pools
HU01987 SVC Critical During SAN fabric power maintenance a cluster may breech resource limits, on the remaining node to node links, resulting in system-wide lease expiry (show details)
Symptom Loss of Access to Data
Environment SVC model SV1 systems running v8.2.1 or later
Trigger Remove a connected SAN switch
Workaround If only two local ports are in use due to zoning and masking for remote copy, and one fabric needs maintenance: Stop partnership; Change port masking - allow node to node comms to use the ports that were remote ports; Perform SAN maintenance
8.2.1.4 Reliability Availability Serviceability
HU02000 All Critical Data Reduction Pools may go offline due to a timing issue in metadata handling (show details)
Symptom Offline Volumes
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.4 Data Reduction Pools
HU02013 All Critical A race condition between the extent invalidation and destruction in the garbage collection process may cause a node warmstart with the possibility of offline volumes (show details)
Symptom Offline Volumes
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.4 Data Reduction Pools
HU02025 All Critical An issue with metadata handling, where a pool has been taken offline, may lead to an out of space condition in that pool preventing its return to operation (show details)
Symptom Offline Volumes
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.4 Data Reduction Pools
HU01886 All High Importance The Unmap function can leave volume extents, that have not been freed, preventing managed disk and pool removal (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
8.2.1.4 SCSI Unmap
HU01902 V5000, V7000 High Importance During an upgrade, an issue with VPD migration, can cause a timeout leading to a stalled upgrade (show details)
Symptom Loss of Redundancy
Environment Storwize systems
Trigger Upgrade
Workaround None
8.2.1.4 System Update
HU01925 FS9100 High Importance Systems will incorrectly report offline and unresponsive NVMe drives after an I/O group outage. These errors will fail to auto-fix and must be manually marked as fixed (show details)
Symptom None
Environment FlashSystem 9100 family systems
Trigger Both canisters in I/O group go offline
Workaround None
8.2.1.4 System Monitoring
HU01930 FS9100 High Importance Certain types of FlashCore Module (FCM) failure may not result in a call home, delaying the shipment of a replacement (show details)
Symptom None
Environment FlashSystem 9100 family systems
Trigger None
Workaround None
8.2.1.4 Drives
HU01937 FS9100, V7000 High Importance DRAID copy-back operation can overload NVMe drives resulting in high I/O latency (show details)
Symptom Performance
Environment Systems with NVMe drives using DRAID
Trigger None
Workaround None
8.2.1.4 Distributed RAID, Drives
HU01939 FS9100, V7000 High Importance After replacing a canister, and attempting to bring the new canister into the cluster, it may remain offline (show details)
Symptom Loss of Redundancy
Environment FlashSystem 9100 family and Storwize V7000 Gen 3 systems running v8.2.1.0 or later
Trigger Using the 'satask chbootdrive -replacecanister' command during node replacement
Workaround Replace both canister boot drives at the same time with new drives
8.2.1.4 Reliability Availability Serviceability
HU01941 All High Importance After upgrading the system to v8.2, or later, when expanding a mirrored volume, the formatting of additional space may become stalled (show details)
Symptom None
Environment Systems upgrading from v8.1.3 or earlier to v8.2.0 or later
Trigger Single or mirrored copy volume exists on system while running v8.1.3.x or earlier Cluster is upgraded from 813 or earlier to 820 or later No extent migrations or changes in host mapping for volume Add copy to volume to make it a volume mirror No extent migrations or changes in host mapping for volume Expand volume size Original volume will not progress format whilst copy will complete Also where both copies of the mirrored volume existed, before the upgrade from v8.1.3.x or earlier, both can have stuck formatting
Workaround None
8.2.1.4 Volume Mirroring
HU01944 All High Importance Proactive host failover not waiting for 25 seconds before allowing nodes to go offline during upgrades or maintenance (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
8.2.1.4 Reliability Availability Serviceability
HU01945 All High Importance Systems with Flash Core Modules are unable to upgrade the firmware for those drives (show details)
Symptom None
Environment Systems using IBM Flash Core Modules
Trigger FCM firmware upgrade
Workaround None
8.2.1.4 Drives
HU01971 FS9100, V7000 High Importance Spurious DIMM over-temperature errors may cause a node to go offline with node error 528 (show details)
Symptom Loss of Redundancy
Environment FlashSystem 9100 family and Storwize V7000 Gen 3 systems
Trigger None
Workaround None
8.2.1.4 Reliability Availability Serviceability
HU01972 All High Importance When an array is in a quiescing state, for example where a member has been deleted, I/O may become pended leading to multiple warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment All systems
Trigger Delete an array member using "charraymember -used unused" command
Workaround None
8.2.1.4 Distributed RAID, RAID
HU00744 All Suggested Single node warmstart due to an accounting issue within the cache component (show details)
Symptom Single Node Warmstart
Environment Systems running v7.3 or later
Trigger None
Workaround None
8.2.1.4 Cache
HU01485 SVC Suggested When a SV1 node is started, with only one PSU powered, powering up the other PSU will not extinguish the Power Fault LED.Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed (show details)
Symptom None
Environment SVC systems using SV1 model nodes
Trigger Power up node with only one PSU powered. Power Fault LED is lit. Power up other PSU. Power Fault LED remains lit.
Workaround Ensure both PSUs are powered before starting node
8.2.1.4 System Monitoring
HU01659 SVC Suggested Node Fault LED can be seen to flash in the absence of an error condition.Note: To apply this fix (in new BMC firmware) each node will need to be power cycled (i.e. remove AC power and battery), one at a time, after the upgrade has completed (show details)
Symptom None
Environment SVC systems using SV1 model nodes
Trigger None
Workaround None
8.2.1.4 System Monitoring
HU01857 All Suggested Improved validation of user input in GUI (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
8.2.1.4 Graphical User Interface
HU01860 All Suggested During garbage collection the flushing of extents may become stuck leading to a timeout and a single node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.4 Data Reduction Pools
HU01869 All Suggested Volume copy deletion, in a Data Reduction Pool, triggered by rmvdiskcopy rmvolumecopy or addvdiskcopy -autodelete (or similar) may become stalled with the copy being left in deleting status (show details)
Symptom None
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.4 Data Reduction Pools
HU01911 All Suggested The System Overview screen, in the GUI, may display nodes in the wrong site (show details)
Symptom None
Environment Systems in a stretched cluster or HyperSwap topology
Trigger None
Workaround None
8.2.1.4 Graphical User Interface
HU01912 All Suggested Systems with iSCSI-attached controllers may see node warmstarts due to I/O request timeouts (show details)
Symptom Single Node Warmstart
Environment Systems running v8.2.1.0 or later with iSCSI-attached controllers
Trigger None
Workaround None
8.2.1.4 Backend Storage
HU01915 & IT28654 All Suggested Systems, with encryption enabled, that are using key servers to manage encryption keys, may fail to connect to the key servers if the servers SSL certificates are part of a chain of trust (show details)
Symptom None
Environment Systems with encryption enabled
Trigger None
Workaround None
8.2.1.4 Encryption
HU01916 All Suggested The GUI Dashboard and the CLI lssystem command report physical capacity incorrectly (show details)
Symptom None
Environment Systems running v8.1 or later
Trigger Upgrading from v8.1 or later
Workaround lsmdisk can continue to be used to provide accurate reporting
8.2.1.4 Command Line Interface, Graphical User Interface
HU01926 SVC, V7000 Suggested When a node, with 32GB of RAM, is upgraded to v8.2.1 it may experience a warmstart resulting in a failed upgrade (show details)
Symptom None
Environment Systems with nodes that have 32GB RAM
Trigger Upgrade to v8.2.1
Workaround None
8.2.1.4 System Update
HU01929 FS9100, V7000 Suggested Drive fault type 3 (error code 1686) may be seen in the Event Log for empty slots (show details)
Symptom None
Environment FlashSystem 9100 family and Storwize V7000 Gen 3 systems
Trigger None
Workaround None
8.2.1.4 System Monitoring
HU01959 All Suggested An timing window issue in the Thin Provisioning component can cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using FlashCopy and Thin Provisioning
Trigger None
Workaround None
8.2.1.4 FlashCopy, Thin Provisioning
HU01961 V5000, V7000 Suggested A hardware issue can provoke the system to repeatedly try to collect a statesave, from the enclosure management firmware, causing 1048 errors in the Event Log (show details)
Symptom None
Environment Storwize systems
Trigger None
Workaround Replace the canister
8.2.1.4 System Monitoring
HU01962 All Suggested When Call Home servers return an invalid message it can be incorrectly reported as an error 3201 in the Event Log (show details)
Symptom None
Environment Systems with Call Home configured
Trigger None
Workaround None
8.2.1.4 System Monitoring
HU01976 All Suggested A new MDisk array may not be encrypted even though encryption is enabled on the system (show details)
Symptom None
Environment Systems running v8.2 or later using encryption without an I/O Group id of 0
Trigger None
Workaround None
8.2.1.4 Encryption
HU02001 All Suggested During a system upgrade an issue in callhome may cause a node warmstart stalling the upgrade (show details)
Symptom Single Node Warmstart
Environment Systems running v8.1 or earlier
Trigger Upgrade system to v8.2 or later
Workaround Before a system upgrade: Remove email servers; Disable cloud call home
8.2.1.4 System Monitoring
HU02002 All Suggested On busy systems, diagnostic data collection may not complete correctly producing livedumps with missing pages (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
8.2.1.4 Support Data Collection
HU02019 All Suggested When the master and auxiliary volumes, in a relationship, have the same name it is not possible, in the GUI, to determine which is master or auxiliary (show details)
Symptom None
Environment Systems using remote copy
Trigger Identical naming of related master and auxiliary volumes
Workaround Use distinguishable names for remote copy master and auxiliary volumes
8.2.1.4 Graphical User Interface
HU02166 All Suggested A timing window issue, in RAID code that handles recovery after a drive has been taken out of sync, due to a slow I/O, can cause a single node warmstart (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
8.2.1.4 RAID
IT28433 All Suggested Timing window issue in the Data Reduction Pool rehoming component can cause a single node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.4 Data Reduction Pools
IT28728 All Suggested Email alerts will not work where the mail server does not allow unqualified client host names (show details)
Symptom Configuration
Environment Systems running v8.2 or later
Trigger None
Workaround None
8.2.1.4 System Monitoring
HU01932 All Critical When a rmvdisk command initiates a Data Reduction Pool rehoming process any I/O to the removed volume may cause multiple warmstarts leading to a loss of access (show details)
Symptom Loss of Access to Data
Environment Systems running v8.2.1.1 using Deduplication
Trigger rmvdisk command will initiate a Data Reduction Pool rehoming process
Workaround None
8.2.1.2 Deduplication
HU01920 All Critical An issue in the garbage collection process can cause node warmstarts and offline pools (show details)
Symptom Offline Volumes
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.1 Data Reduction Pools
HU01492 & HU02024 SVC, V5000, V7000 HIPER All ports of a 16Gb HBA can be affected when a single port is congested. This can lead to lease expiries if all ports used for inter-node communication are on the same FC adapter (show details)
Symptom Loss of Access to Data
Environment Systems using 16Gb HBAs
Trigger All ports used for inter-node communication are on the same FC adapter and a port on that adapter experiences congestion
Workaround Separate inter-node traffic so that multiple adapters are used
8.2.1.0 Reliability Availability Serviceability
HU01617 All HIPER Due to a timing window issue, stopping a FlashCopy mapping, with the -autodelete option, may result in a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems using FlashCopy
Trigger None
Workaround Avoid stopping FlashCopy mappings with the -autodelete option
8.2.1.0 FlashCopy
HU01828 All HIPER Node warmstarts may occur during deletion of deduplicated volumes due to a timing-related issue (show details)
Symptom Loss of Access to Data
Environment Systems using deduplicated volume copies
Trigger Deleting a deduplication volume copy
Workaround Do not delete deduplicated volume copies
8.2.1.0 Deduplication
HU01851 All HIPER When a deduplicated volume is deleted there may be multiple node warmstarts and offline pools (show details)
Symptom Loss of Access to Data
Environment Systems running v8.1.3 or later using Deduplication
Trigger Delete a deduplicated volume
Workaround None
8.2.1.0 Data Reduction Pools, Deduplication
HU01873 All HIPER Deleting a volume, in a Data Reduction Pool, while volume protection is enabled and when the volume was not explicitly unmapped, before deletion, may result in simultaneous node warmstarts. For more details refer to this Flash (show details)
Symptom Loss of Access to Data
Environment Systems using Data Reduction Pools
Trigger Delete volume in Data Reduction Pool while volume protection is enabled
Workaround Either: Disable volume protection; or Remove host mappings before deleting a volume. If using scripts, modify them to unmap volumes before deletion.
8.2.1.0 Data Reduction Pools
HU01906 FS9100 HIPER Low-level hardware errors may not be recovered correctly, causing a canister to reboot. If multiple canisters reboot, this may result in loss of access to data (show details)
Symptom Multiple Node Warmstarts
Environment FlashSystem 9100 family systems
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU01913 All HIPER A timing window issue in the DRAID6 rebuild process can cause node warmstarts with the possibility of a loss of access (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID
Trigger None
Workaround None
8.2.1.0 Distributed RAID
HU01743 All Critical Where hosts are directly attached a mishandling of the login process, by the fabric controller, may result in dual node warmstarts (show details)
Symptom Loss of Access to Data
Environment Systems with direct-attached hosts
Trigger None
Workaround None
8.2.1.0 Hosts
HU01758 All Critical After an unexpected power loss, all nodes, in a cluster, may warmstart repeatedly, necessitating a Tier 3 recovery (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger Power outage
Workaround None
8.2.1.0 RAID
HU01799 All Critical Timing window issue can affect operation of the HyperSwap addvolumecopy command causing all nodes to warmstart (show details)
Symptom Loss of Access to Data
Environment Systems using HyperSwap
Trigger None
Workaround Ensuring that the volume is not in a FlashCopy mapping will allow the use of the addvolumecopy. Alternatively the addvolumecopy command combines a number of CLI commands for convenience. Use these CLI commands individually to achieve the same outcome
8.2.1.0 HyperSwap
HU01825 All Critical Invoking a chrcrelationship command when one of the relationships in a consistency group is running in the opposite direction to the others may cause a node warmstart followed by a Tier 2 recovery (show details)
Symptom Loss of Access to Data
Environment Systems using HyperSwap
Trigger A relationship running in one direction is added to a consistency group running in the other direction whilst one of the FlashCopy maps associated with the HyperSwap relationship is still stopping/cleaning
Workaround Do not add a relationship to a consistency group if they are running in opposite directions (i.e. the Primary of the consistency group and the Primary of the relationship are on different sites); Do not add a relationship to a consistency group if the relationship still has one of its FlashCopy maps in the stopping state. The clean progress needs to reach 100 percent before the relationship can be safely added.
8.2.1.0 FlashCopy
HU01833 All Critical If both nodes, in an I/O group, start up together a timing window issue may occur, that would prevent them running garbage collection, leading to a related Data Reduction Pool running out of space (show details)
Symptom Offline Volumes
Environment Systems using Data Reduction Pools
Trigger Start both nodes in an I/O group at the same time
Workaround Ensure nodes in an I/O group start one at a time
8.2.1.0 Data Reduction Pools
HU01845 All Critical If the execution of a rmvdisk -force command, for the FlashCopy target volume in a GMCV relationship, coincides with the start of a GMCV cycle all nodes may warmstart (show details)
Symptom Loss of Access to Data
Environment Systems running v7.7.1 or later using Global Mirror with Change Volumes
Trigger rmvdisk -force command execution coincides with the start of a GMCV cycle
Workaround Remove the related FC map before issuing a rmvdisk command without -force
8.2.1.0 Global Mirror With Change Volumes
HU01847 All Critical FlashCopy handling of medium errors across a number of drives on backend controllers may lead to multiple node warmstarts (show details)
Symptom Loss of Access to Data
Environment Systems running v7.8.1 or later using FlashCopy
Trigger None
Workaround None
8.2.1.0 FlashCopy
HU01848 All Critical During an upgrade, systems with a large AIX VIOS setup may have multiple node warmstarts with the possibility of a loss of access to data (show details)
Symptom Loss of Access to Data
Environment Systems presenting storage to large IBM AIX VIOS configurations
Trigger None
Workaround None
8.2.1.0 System Update
HU01850 All Critical When the last deduplication-enabled volume copy in a Data Reduction Pool is deleted the pool may go offline temporarily (show details)
Symptom Loss of Access to Data
Environment Systems using Data Reduction Pools with deduplicated volume copies
Trigger Delete last deduplication-enabled volume copy in a Data Reduction Pool
Workaround If a Data Reduction Pool contains volumes with deduplication enabled keep at least one of those volumes in the pool
8.2.1.0 Data Reduction Pools, Deduplication
HU01855 All Critical Clusters using Data Reduction Pools can experience multiple warmstarts on all nodes putting them in a service state (show details)
Symptom Loss of Access to Data
Environment Systems running v8.1.2 or later using Data Reduction Pools
Trigger None
Workaround None
8.2.1.0 Data Reduction Pools
HU01862 All Critical When a Data Reduction Pool is removed, and the -force option is specified, there may be a temporary loss of access (show details)
Symptom Loss of Access to Data
Environment Systems using Data Reduction Pools
Trigger Remove a Data Reduction Pool with the -force option
Workaround Do not use -force option when removing a Data Reduction Pool
8.2.1.0 Data Reduction Pools
HU01876 All Critical Where systems are connected to controllers, that have FC ports that are capable of acting as initiators and targets, when NPIV is enabled then node warmstarts can occur (show details)
Symptom Loss of Access to Data
Environment Systems, with NPIV enabled, attached to host ports that can act as SCSI initiators and targets
Trigger Zone host initiator and target ports in with the target port WWPN then enable NPIV
Workaround Unzone host or disable NPIV
8.2.1.0 Backend Storage
HU01878 All Critical During an upgrade from v7.8.1 or earlier to v8.1.3 or later if an MDisk goes offline then at completion all volumes may go offline (show details)
Symptom Offline Volumes
Environment Systems running v7.8.1 or earlier
Trigger MDisk goes offline during an upgrade to v8.1.3 or later
Workaround None
8.2.1.0 System Update
HU01885 All Critical As writes are made to a Data Reduction Pool it is necessary to allocate new physical capacity. Under unusual circumstances it is possible for the handling of an expansion request to stall further I/O leading to node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.0 Data Reduction Pools
HU01901 V7000 Critical Enclosure management firmware, in an expansion enclosure, will reset a canister after a certain number of discovery requests have been received, from the controller, for that canister. It is possible simultaneous resets may occur in adjacent canisters causing a temporary loss of access to data (show details)
Symptom Loss of Access to Data
Environment Storwize expansion enclosures running v8.2.0
Trigger One million discovery requests received from attached controller
Workaround Proactively reset individual expansion canisters at convenient times
8.2.1.0 Reliability Availability Serviceability
HU01957 All Critical Due to an issue in Data Reduction Pools, when the system attempts an upgrade, there may be node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems using Data Reduction Pools
Trigger Initiate system upgrade
Workaround None
8.2.1.0 Data Reduction Pools, System Update
HU01965 All Critical A timing window issue in the deduplication component can lead to I/O timeouts, and a node warmstart, with the possibility of an offline MDisk group (show details)
Symptom Offline Volumes
Environment Systems running v8.1.3 or later using Deduplication
Trigger None
Workaround None
8.2.1.0 Deduplication
HU02042 All Critical An issue in the handling of metadata, after a Data Reduction Pool recovery operation, can lead to repeated node warmstarts, putting an I/O group into a service state (show details)
Symptom Loss of Access to Data
Environment Systems using Data Reduction Pools
Trigger T3 recovery
Workaround None
8.2.1.0 Data Reduction Pools
IT25850 All Critical I/O performance may be adversely affected towards the end of DRAID rebuilds. For some systems there may be multiple warmstarts leading to a loss of access (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID
Trigger None
Workaround None
8.2.1.0 Distributed RAID
IT27460 All Critical Lease expiry can occur between local nodes when remote connection is lost, due to the mishandling of messaging credits (show details)
Symptom Loss of Access to Data
Environment All systems
Trigger None
Workaround Use four ports for local to local node communications, on at least two separate fibre channel adapters per node. Port mask so that all four are usable. Use a different fibre channel adapter than the above two adapters for remote port communications. If there are issues with the FCIP tunnel, temporarily block that until it is fixed.
8.2.1.0 Reliability Availability Serviceability
IT29040 All Critical Occasionally a DRAID rebuild, with drives of 8TB or more, can encounter an issue which causes node warmstarts and potential loss of access (show details)
Symptom Loss of Access to Data
Environment Systems using DRAID with drives of 8TB or more
Trigger None
Workaround None
8.2.1.0 Distributed RAID, RAID
IT29853 V5000 Critical After upgrading to v8.1.1, or later, V5000 Gen 2 systems, with Gen 1 expansion enclosures, may experience multiple node warmstarts leading to a loss of access (show details)
Symptom Loss of Access to Data
Environment Storwize V5000 Gen 2 systems with Gen 1 expansion enclosures
Trigger Upgrade to v8.1.1 or later
Workaround None
8.2.1.0 System Update
HU01507 All High Importance Until the initial synchronisation process completes, high system latency may be experienced when a volume is created with two compressed copies or when space-efficient copy is added to a volume with an existing compressed copy (show details)
Symptom Performance
Environment All systems
Trigger Create a volume with two compressed copies or add a space-efficient copy to a volume with an existing compressed copy
Workaround Avoid: creating a new volume with two compressed copies; adding a SE volume copy to a volume that already possesses a compressed copy
8.2.1.0 Volume Mirroring
HU01661 All High Importance A cache-protection mechanism flag setting can become stuck leading to repeated stops of consistency group synchronisation (show details)
Symptom Loss of Redundancy
Environment Systems running v7.6 or later using remote copy
Trigger None
Workaround None
8.2.1.0 HyperSwap
HU01733 All High Importance Canister information, for the High Density Expansion Enclosure, may be incorrectly reported (show details)
Symptom Loss of Redundancy
Environment Systems using the High Density Expansion Enclosure (92F)
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU01761 All High Importance Entering multiple addmdisk commands, in rapid succession, to more than one storage pool, may cause node warmstarts (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v8.1 or later with two or more storage pools
Trigger Run multiple addmdisk commands to more than one storage pool at the same time
Workaround Paced addmdisk commands to one storage pool at a time
8.2.1.0 Backend Storage
HU01797 All High Importance Hitachi G1500 backend controllers may exhibit higher than expected latency (show details)
Symptom Performance
Environment Systems with Hitachi G1500 backend controllers
Trigger None
Workaround None
8.2.1.0 Backend Storage
HU01810 All High Importance Deleting volumes, or using FlashCopy/Global Mirror with Change Volumes, in a Data Reduction Pool, may impact the performance of other volumes in the pool (show details)
Symptom Performance
Environment Systems using Data Reduction Pools
Trigger Volume deletion or FlashCopy/GMCV
Workaround None
8.2.1.0 Data Reduction Pools
HU01837 All High Importance In systems where a vVols metadata volume has been created an upgrade to v8.1.3 or later will cause a node warmstart stalling the upgrade (show details)
Symptom Loss of Redundancy
Environment Systems running v8.1.0, v8.1.1 or v8.1.2 that are providing vVols
Trigger Upgrading to v8.1.3 or later
Workaround Contact support if system is running v8.1.2. Otherwise this workaround can be used: Use svcinfo lsmetadatavdisk to find the volume id; Create a new volume copy in the same MDisk group - svctask addvdiskcopy -mdiskgrp X -autodelete <vdisk_id>; Wait until lsvdisksyncprogress no longer shows a mirror in progress; Upgrade
8.2.1.0 System Update, vVols
HU01839 All High Importance Where a VMware host is being served volumes, from two different controllers, and an issue, on one controller, causes the related volumes to be taken offline then I/O performance, for the volumes from the other controller, will be adversely affected (show details)
Symptom Performance
Environment Systems running v7.5 or later presenting volumes to VMware hosts, from more than one back-end controller
Trigger Issue on back-end controller takes volumes offline
Workaround None
8.2.1.0 Hosts
HU01842 All High Importance Bursts of I/O to Read-Intensive Drives can be interpreted as dropped frames against the resident slots, leading to redundant drives being incorrectly failed (show details)
Symptom Loss of Redundancy
Environment Systems with Read-Intensive Drives
Trigger None
Workaround None
8.2.1.0 Drives
HU01846 SVC High Importance Silent battery discharge condition will unexpectedly take a SVC node offline putting it into a 572 service state (show details)
Symptom Loss of Redundancy
Environment SVC systems using DH8 & SV1 model nodes
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU01852 All High Importance The garbage collection rate can lead to Data Reduction Pools running out of space even though reclaimable capacity is available (show details)
Symptom None
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.0 Data Reduction Pools
HU01858 All High Importance Total used capacity of a Data Reduction Pool within a single I/O group is limited to 256TB. Garbage collection does not correctly recognise this limit. This may lead to a pool running out of free capacity and going offline (show details)
Symptom None
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.0 Data Reduction Pools
HU01881 FS9100 High Importance An issue within the compression card in FS9100 systems can result in the card being incorrectly flagged as failed leading to warmstarts (show details)
Symptom Loss of Redundancy
Environment FS9100 systems
Trigger None
Workaround None
8.2.1.0 Compression
HU01883 All High Importance Config node processes may consume all available memory, leading to node warmstarts. This can be caused, for example, by large numbers of concurrent SSH connections being opened (show details)
Symptom Multiple Node Warmstarts
Environment Systems running v8.1.3 or later
Trigger Many concurrent SSH connections
Workaround Reduce number of SSH connections
8.2.1.0 Reliability Availability Serviceability
HU01907 SVC High Importance An issue in the handling of the power cable sense registers can cause a node to be put into service state with a 560 error (show details)
Symptom Loss of Redundancy
Environment SVC systems using SV1 model nodes
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU01934 FS9100 High Importance An issue in the handling of faulty canister components can lead to multiple node warmstarts for that canister (show details)
Symptom Multiple Node Warmstarts
Environment FlashSystem 9100 family systems
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU00921 All Suggested A node warmstart may occur when an MDisk state change gives rise to duplicate discovery processes (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
8.2.1.0
HU01276 All Suggested An issue in the handling of debug data from the FC adapter can cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU01523 All Suggested An issue with FC adapter initialisation can lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU01564 All Suggested FlashCopy maps cleaning process is not monitoring the grains correctly which may cause FlashCopy maps to not stop (show details)
Symptom None
Environment Systems using FlashCopy
Trigger None
Workaround None
8.2.1.0 FlashCopy
HU01571 All Suggested An upgrade can become stalled due to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems undergoing a code upgrade
Trigger None
Workaround None
8.2.1.0 System Update
HU01657 SVC, V5000, V7000 Suggested The 16Gb FC HBA firmware may experience an issue, with the detection of unresponsive links, leading to a single node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU01667 All Suggested A timing-window issue, in the remote copy component, may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using remote copy
Trigger None
Workaround None
8.2.1.0 Global Mirror, Global Mirror With Change Volumes, Metro Mirror
HU01719 All Suggested Node warmstart due to a parity error in the HBA driver firmware (show details)
Symptom Single Node Warmstart
Environment Systems running v7.6 and later using 16Gb HBAs
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU01737 All Suggested On the Update System screen, for Test Only, if a valid code image is selected, in the Run Update Test Utility dialog, then clicking the Test button will initiate a system update (show details)
Symptom None
Environment All systems
Trigger Select a valid code image in the "Run Update Test Utility" dialog and click "Test" button
Workaround Do not select a valid code image in the "Test utility" field of the "Run Update Test Utility" dialog
8.2.1.0 System Update
HU01760 All Suggested FlashCopy map progress appears to be stuck at zero percent (show details)
Symptom None
Environment Systems using FlashCopy
Trigger None
Workaround None
8.2.1.0 FlashCopy
HU01765 All Suggested Node warmstart may occur when there is a delay to I/O at the secondary site (show details)
Symptom Single Node Warmstart
Environment Systems using remote copy
Trigger None
Workaround None
8.2.1.0 Global Mirror, Global Mirror With Change Volumes, Metro Mirror
HU01772 All Suggested The mail queue may become blocked preventing the transmission of event log messages (show details)
Symptom None
Environment All systems
Trigger None
Workaround None
8.2.1.0 System Monitoring
HU01784 All Suggested If a cluster using IP quorum experiences a site outage, the IP quorum device may become invalid. Restarting the quorum application will resolve the issue (show details)
Symptom Configuration
Environment Systems using Hyperswap or enhanced stretched cluster
Trigger Outage on all nodes at a single site
Workaround Restart the quorum application after the site outage
8.2.1.0 HyperSwap, Quorum
HU01786 All Suggested An issue in the monitoring of SSD write endurance can result in false 1215/2560 errors in the Event Log (show details)
Symptom None
Environment Systems running v7.7.1 or later with SSDs
Trigger None
Workaround None
8.2.1.0 Drives
HU01791 All Suggested Using the chhost command will remove stored CHAP secrets (show details)
Symptom Configuration
Environment Systems using iSCSI
Trigger Run the "chhost -gui -name <host name> <host id>" command after configuring CHAP secret
Workaround Set the CHAP secret whenever changing the host name
8.2.1.0 iSCSI
HU01807 All Suggested The lsfabric command may show incorrect local node id and local node name for some Fibre Channel logins (show details)
Symptom None
Environment All systems
Trigger None
Workaround Use the local WWPN and reference the node in lsportfc to get the correct information
8.2.1.0 Command Line Interface
HU01811 All Suggested DRAID rebuilds, for large (>10TB) drives, may require lengthy metadata processing leading to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using DRAID
Trigger None
Workaround None
8.2.1.0 Distributed RAID
HU01815 All Suggested In Data Reduction Pools, volume size is limited to 96TB (show details)
Symptom None
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.0 Data Reduction Pools
HU01817 All Suggested Volumes used for vVols metadata or cloud backup, that are associated with a FlashCopy mapping, cannot be included in any further FlashCopy mappings (show details)
Symptom Configuration
Environment Systems using vVols or TCT
Trigger None
Workaround None
8.2.1.0 FlashCopy
HU01821 SVC Suggested An attempt to upgrade a two-node enhanced stretched cluster fails due to incorrect volume dependencies (show details)
Symptom None
Environment Systems configured as a two-node enhanced stretched cluster that are using Data Reduction Pools
Trigger Upgrade
Workaround Revert cluster to standard topology and remove site settings from nodes and controllers for the duration of the upgrade
8.2.1.0 Data Reduction Pools, System Update
HU01832 All Suggested Creation and distribution of the config file may cause an out-of-memory condition, leading to a node warmstart (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU01849 All Suggested An excessive number of SSH sessions may lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger Initiate a large number of SSH sessions (e.g. one session every 5 seconds)
Workaround Avoid initiating excessive numbers of SSH sessions
8.2.1.0 System Monitoring
HU01856 All Suggested A garbage collection process can time out waiting for an event in the partner node resulting in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.0 Data Reduction Pools
HU01863 All Suggested In rare circumstances, a drive replacement may result in a ghost drive (i.e. a drive with the same ID as the replaced drive stuck in a permanently offline state) (show details)
Symptom None
Environment All systems
Trigger Drive replacement
Workaround None
8.2.1.0 Drives
HU01871 All Suggested An issue with bitmap synchronisation can lead to a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Data Reduction Pools
Trigger None
Workaround None
8.2.1.0 Data Reduction Pools
HU01879 All Suggested Latency induced by DWDM inter-site links may result in a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using DWDM inter-site links
Trigger None
Workaround None
8.2.1.0
HU01893 FS9100, SVC, V7000 Suggested Excessive reporting frequency of NVMe drive diagnostics generates large numbers of callhome events (show details)
Symptom None
Environment Systems running v8.2 or later using NVMe drives
Trigger None
Workaround None
8.2.1.0 Drives
HU01895 All Suggested Where a banner has been created, without a new line at the end, any subsequent T4 recovery will fail (show details)
Symptom None
Environment Systems using DRAID with a banner set
Trigger Initiate a T4 recovery
Workaround Modify svc.config.backup.xml - add new line at the end of banner stanza. Before the </object> tag
8.2.1.0 Distributed RAID
HU01981 All Suggested Although an issue, in the HBA firmware, is handled correctly it can still cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
HU02028 All Suggested An issue, with timer cancellation, in the Remote Copy component may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using Remote Copy
Trigger None
Workaround None
8.2.1.0 Global Mirror, Global Mirror With Change Volumes, Metro Mirror
IT19561 All Suggested An issue with register clearance in the FC driver code may cause a node warmstart (show details)
Symptom Single Node Warmstart
Environment Systems using 16Gb HBAs
Trigger None
Workaround None
8.2.1.0 Reliability Availability Serviceability
IT25457 All Suggested Attempting to remove a copy of a volume, which has at least one image mode copy and at least one thin/compressed copy, in a Data Reduction Pool will always fail with a CMMVC8971E error (show details)
Symptom None
Environment Systems using Data Reduction Pools
Trigger Try to remove a copy of a volume, which has at least one image mode copy and at least one thin/compressed copy, in a Data Reduction Pool
Workaround Use svctask splitvdiskcopy to create a separate volume from the copy that should be deleted. This leaves the original volume with a single copy and creates a new volume from the copy that was split off. Then remove the newly created volume.
8.2.1.0 Data Reduction Pools
IT25970 All Suggested After a FlashCopy consistency group is started a node may warmstart (show details)
Symptom Single Node Warmstart
Environment All systems
Trigger None
Workaround None
8.2.1.0 FlashCopy
IT26049 All Suggested An issue with CPU scheduling may cause the GUI to respond slowly (show details)
Symptom None
Environment Systems running v7.8 or later
Trigger None
Workaround None
8.2.1.0 Graphical User Interface

4. Useful Links

Description Link
Support Websites
Update Matrices, including detailed build version
Support Information pages providing links to the following information:
  • Interoperability information
  • Product documentation
  • Limitations and restrictions, including maximum configuration limits
Spectrum Virtualize Family of Products Inter-System Metro Mirror and Global Mirror Compatibility Cross Reference
Software Upgrade Test Utility
Software Upgrade Planning