Note: The 8.3.1.0 release is only provided, pre-installed, on new systems and will not be available on IBM Fix Central.
This document was last updated on 28 October 2024.
Note. Detailed build version numbers are included in the Update Matrices in the Useful Links section.Details | Introduced |
---|---|
Due to security improvements, when a customer upgrades to v8.3.1, for the first time, an event 3330 ("The superuser password must be changed") will be seen in the Event Log. The prior superuser password will continue to be accepted until it is changed. Customers can delay changing the superuser password until it is convenient. The action of changing the password will enable the improved security. If the customer wishes, the superuser password can be set to the same password. |
8.3.1.0 |
Customers using Spectrum Control v5.3.2, or earlier, may notice some discrepancies between attributes displayed and related attributes on the Spectrum Virtualize GUI. This issue will be resolved by a future release of Spectrum Control. |
8.3.1.0 |
Under some circumstances, the "Monitoring > System" GUI screen may not load completely. This is a known issue that will be lifted in a future PTF. |
8.3.1.0 |
Remote Support Assistance (RSA) does not work after changing SSH node method from password to key on Amazon WebServices (AWS). This is a known issue that may be lifted in a future PTF. |
8.3.0.0 |
Customers using iSER attached hosts, with Mellanox 25G adapters, should be aware that IPv6 sessions will not failover, for example, during a cluster upgrade. This is a known issue that may be lifted in a future PTF. |
8.3.0.0 |
Due to an issue in the automatic upgrade process, customers with 32GB DH8 systems, running v8.2.1.4 or v8.2.1.5, cannot upgrade to v8.2.1.6 or later. There is a manual upgrade option for these systems. Affected customers, wishing to use this option, should contact IBM Support for assistance quoting APAR HU02032 This is a restriction that may be lifted in a future PTF. |
8.2.1.5 |
Customers upgrading systems with more than 64GB of RAM to v8.1 or later will need to run chnodehw to enable access to the extra memory above 64GB. Under some circumstances it may also be necessary to remove and re-add each node in turn. |
8.1.0.0 |
Systems, with NPIV enabled, presenting storage to SUSE Linux Enterprise Server (SLES) or Red Hat Enterprise Linux (RHEL) hosts running the ibmvfc driver on IBM Power can experience path loss or read-only file system events. This is cause by issues within the ibmvfc driver and VIOS code. Refer to this troubleshooting page for more information. |
n/a |
If an update stalls or fails then contact IBM Support for further assistance | n/a |
The following restrictions were valid but have now been lifted | |
Reoccurring node warmstarts on systems with DRP that have been upgraded to 8.3.1.7 or 8.3.1.8. This issue has been resolved by APAR HU02485. |
8.3.1.9 |
Systems with 92-drive expansion enclosures can experience fan failures after the software upgrade. This is because system firmware has been updated to detect fans that are not operating correctly. The root cause of this issue was determined to be faulty fans and not over-zealous firmware |
8.3.1.4 |
There is an issue where the Spectrum Virtualize REST API output for action commands is in the standard CLI format and not JSON. This will result in failures of the IBM Ansible modules when creating objects. This issue has been resolved in PTF v8.3.1.6. |
8.3.1.3 |
Due to an issue in IP Replication, customers using this feature should not upgrade to v8.3.1.3 or later. This issue has been resolved by APARs HU02340 in PTF v8.3.1.4. |
8.3.1.3 |
Configuration node warmstart will occur if mkhostcluster is run with -ignoreseedvolume and the ignored volumes have an id greater than 256. This issue has been resolved by APARs HU02303 & HU02305 in PTF v8.3.1.3. |
8.3.1.0 |
When upgrading from v8.2.1, or earlier, to v8.3.0, or later, the CLI and GUI may incorrectly show all hosts offline. Using the CLI svcinfo lshost <host id number> command for individual hosts will show the host is actually online. This issue has been resolved by APAR HU02281 in PTF v8.3.1.3. |
8.3.1.0 |
Customers using a Cisco Fibre Channel network, that is running NX-OS v8.3(2), or earlier, must not upgrade to Spectrum Virtualize v8.3.1 or later. Please refer to the Flash for details. This issue has been resolved by APAR HU02182 in PTF v8.3.1.2. |
8.3.1.0 |
Customers with systems running v8.3.1.0 or later must avoid expanding Distributed RAID arrays containing FlashCore Modules. Please refer to the Flash for details. This issue has been resolved in PTF v8.3.1.2. |
8.3.1.0 |
Customers with systems running v8.3.0.0, or earlier, using deduplication cannot upgrade to v8.3.0.1, or later, due to APAR HU02162. This issue has been resolved by APAR HU02162 in PTF v8.3.1.3. |
8.3.0.1 |
Validation in the Upload Support Package feature will reject the new case number format in the PMR field. This issue has been resolved by APAR HU02392. |
7.8.1.0 |
A release may contain fixes for security issues, fixes for APARs or both. Consult both tables below to understand the complete set of fixes included in the release.
CVE Identifier |
Link for additional Information |
Resolved in |
---|---|---|
| ||
CVE-2023-21930 | 7065011 | 8.3.1.10 |
CVE-2023-21937 | 7065011 | 8.3.1.108 |
CVE-2023-21938 | 7065011 | 8.3.1.10 |
CVE-2023-43042 | 7064976 | 8.3.1.10 |
CVE-2022-39167 | 6622025 | 8.3.1.9 |
CVE-2018-25032 | 6622021 | 8.3.1.8 |
CVE-2022-0778 | 6622017 | 8.3.1.7 |
CVE-2021-35603 | 6622019 | 8.3.1.7 |
CVE-2021-35550 | 6622019 | 8.3.1.7 |
CVE-2021-38969 | 6584337 | 8.3.1.7 |
CVE-2021-29873 | 6497111 | 8.3.1.6 |
CVE-2020-2781 | 6445063 | 8.3.1.3 |
CVE-2020-13935 | 6445063 | 8.3.1.3 |
CVE-2020-14577 | 6445063 | 8.3.1.3 |
CVE-2020-14578 | 6445063 | 8.3.1.3 |
CVE-2020-14579 | 6445063 | 8.3.1.3 |
CVE-2020-4686 | 6260199 | 8.3.1.2 |
CVE-2019-5544 | 6250889 | 8.3.1.0 |
CVE-2019-2964 | 6250887 | 8.3.1.0 |
CVE-2019-2989 | 6250887 | 8.3.1.0 |
CVE-2018-12404 | 6250885 | 8.3.1.0 |
APAR |
Affected Products |
Severity |
Description |
Resolved in |
Feature Tags |
---|---|---|---|---|---|
|
|
| |||
HU02467 | All | Critical | When one node disappears from the cluster the surviving node can be unable to achieve quorum allegiance in a timely manner causing it to lease expire (show details) | 8.3.1.9 | Quorum |
HU02471 | All | Critical | After starting a FlashCopy map with -restore in a graph with a GMCV secondary disk that was stopped with -access there can be a data integrity issue (show details) | 8.3.1.9 | FlashCopy, Global Mirror With Change Volumes |
HU02561 | All | Critical | If there are a high number of FC mappings sharing the same target, the internal array that is used to track the FC mapping is mishandled, thereby causing it to overrun. This will cause a cluster wide warmstart to occur (show details) | 8.3.1.9 | FlashCopy |
IT41088 | FS5000, FS5100, V5000, V5100 | Critical | Systems with low memory that have a large number of RAID arrays that are resyncing can cause a system to run out of RAID rebuild control blocks (show details) | 8.3.1.9 | RAID |
HU02010 | All | High Importance | A single node warmstart may occur when a drive in a non-distributed RAID array is taken temporarily out-of-sync due to slow performance (show details) | 8.3.1.9 | RAID |
HU02485 | All | High Importance | Reoccurring node warmstarts on systems with DRP that have been upgraded to 8.3.1.7 or 8.3.1.8 (show details) | 8.3.1.9 | Data Reduction Pools, System Update |
IT41835 | All | High Importance | A T2 recovery may occur when a failed drive in the system is replaced with an unsupported drive type (show details) | 8.3.1.9 | Drives |
HU02306 | All | Suggested | An offline host port can still be shown as active in lsfabric and the associated host can be shown as online despite being offline (show details) | 8.3.1.9 | Hosts |
HU02364 | All | Suggested | False 989001 Managed Disk Group space warnings can be generated (show details) | 8.3.1.9 | System Monitoring |
HU02367 | All | Suggested | An issue with how RAID handles drive failures may lead to a node warmstart (show details) | 8.3.1.9 | RAID |
HU02372 | FS9100, SVC, V5000, V5100, V7000 | Suggested | Host SAS port 4 is missing from the GUI view on some systems. (show details) | 8.3.1.9 | Graphical User Interface |
HU02391 | All | Suggested | An issue with how websockets connections are handled can cause the GUI to become unresponsive requiring a restart of the Tomcat server (show details) | 8.3.1.9 | Graphical User Interface |
HU02443 | All | Suggested | An inefficiency in the RAID code that processes requests to free memory can cause the request to timeout leading to a node warmstart (show details) | 8.3.1.9 | RAID |
HU02453 | All | Suggested | It may not be possible to connect to GUI or CLI without a restart of the Tomcat server (show details) | 8.3.1.9 | Command Line Interface, Graphical User Interface |
HU02474 | All | Suggested | An SFP failure can cause a node warmstart (show details) | 8.3.1.9 | Reliability Availability Serviceability |
HU02499 | All | Suggested | A pop up with the message saying 'The server was unable to process the request' may occur due to an invalid time stamp in the file used to provide the pop up reminder (show details) | 8.3.1.9 | Graphical User Interface |
HU02564 | All | Suggested | The 'charraymember' command fails with a degraded DRAID array, even though the syntax of the command is correct (show details) | 8.3.1.9 | Distributed RAID |
HU02593 | All | Suggested | NVMe drive is incorrectly reporting end of life due to flash degradation (show details) | 8.3.1.9 | Drives |
HU02409 | All | Critical | If rmhost command, with -force, is executed for a MS Windows server then an issue in the iSCSI driver can cause the relevant target initiator to become unresponsive (show details) | 8.3.1.7 | Hosts, iSCSI |
HU02410 | SVC | Critical | A timing window issue in the transition to a spare node can cause a cluster-wide Tier 2 recovery (show details) | 8.3.1.7 | Hot Spare Node |
HU02455 | All | Critical | After converting a system from 3-site to 2-site a timing window issue can trigger a cluster tier 2 recovery (show details) | 8.3.1.7 | 3-Site using HyperSwap or Metro Mirror |
HU02343 | All | High Importance | For Huawei Dorado V3 Series backend controllers it is possible that not all available target ports will be utilized. This would reduce the potential IO throughput and can cause high read/write backend queue time on the cluster impacting front end latency for hosts (show details) | 8.3.1.7 | Backend Storage |
HU02460 | All | High Importance | Multiple node warmstarts triggered by ports on the 32G fibre channel adapter failing (show details) | 8.3.1.7 | Hosts |
HU02466 | All | High Importance | An issue in the handling of drive failures can result in multiple node warmstarts (show details) | 8.3.1.7 | RAID |
HU01209 | All | Suggested | It is possible for the Fibre Channel driver to be offered an unsupported length of data resulting in a node warmstart (show details) | 8.3.1.7 | Storage Virtualisation |
HU02433 | FS5000, FS5100, SVC, V5000, V5100, V7000 | Suggested | When a BIOS upgrade occurs excessive tracefile entries can be generated (show details) | 8.3.1.7 | System Update |
HU02451 | All | Suggested | An incorrect IP Quorum lease extension setting can lead to a node warmstart (show details) | 8.3.1.7 | IP Quorum |
IT33996 | All | Suggested | An issue in RAID where unreserved resources fail to be freed up can result in a node warmstart (show details) | 8.3.1.7 | RAID |
HU02296 | All | HIPER | The zero page functionality can become corrupt causing a volume to be initialised with non-zero data (show details) | 8.3.1.6 | Storage Virtualisation |
HU02327 | All | HIPER | Using addvdiskcopy in conjunction with expandvdisk with format may result in the original being overwritten, by the new copy, producing blank copies. For more details refer to this Flash (show details) | 8.3.1.6 | Volume Mirroring |
HU02384 | SVC | HIPER | An inter-node message queue can become stalled, leading to an I/O timeout warmstart, and temporary loss of access (show details) | 8.3.1.6 | Reliability Availability Serviceability |
HU02400 | All | HIPER | A problem in the virtualization component of the system can cause a migration IO to be submitted in an incorrect context resulting in a node warmstart. In some cases it is possible that this IO has been submitted to an incorrect location on the backend, which can cause data corruption of an isolated small area (show details) | 8.3.1.6 | Storage Virtualisation |
HU02418 | All | HIPER | During a DRAID array rebuild data can be written to an incorrect location. For more details refer to this Flash (show details) | 8.3.1.6 | Distributed RAID, RAID |
DT112601 | All | Critical | Deleting image mode mounted source volume while migration is ongoing could trigger Tier 2 recovery (show details) | 8.3.1.6 | Storage Virtualisation |
HU02226 | All | Critical | Due to an issue in DRP a node can repeatedly warmstart whilst rejoining a cluster (show details) | 8.3.1.6 | Data Reduction Pools |
HU02342 | All | Critical | Occasionally when an offline drive returns to online state later than its peers in the same RAID array there can be multiple node warmstarts that send nodes into a service state (show details) | 8.3.1.6 | RAID |
HU02373 | All | Critical | An incorrect compression flag in metadata can take a DRP offline (show details) | 8.3.1.6 | Data Reduction Pools |
HU02393 | All | Critical | Automatic resize of compressed/thin volumes may fail causing warmstarts on both nodes in an I/O group (show details) | 8.3.1.6 | Storage Virtualisation |
HU02397 | All | Critical | A Data Reduction Pool, with deduplication enabled, can retain some stale state after deletion and recreation. This has no immediate effect. However if later on a node goes offline this condition can cause the pool to be taken offline (show details) | 8.3.1.6 | Data Reduction Pools |
HU02401 | All | Critical | EasyTier can move extents between identical mdisks until one runs out of space (show details) | 8.3.1.6 | EasyTier |
HU02406 | All | Critical | An interoperability issue between Cisco NX-OS firmware and the Spectrum Virtualize Fibre Channel driver can cause a node warmstart on NPIV failback (for example during an upgrade) with the potential for a loss of access. For more details refer to this Flash (show details) | 8.3.1.6 | Interoperability |
HU02414 | All | Critical | Under specific sequence and timing of circumstances the garbage collection process can timeout and take a pool offline temporarily (show details) | 8.3.1.6 | Data Reduction Pools |
HU02429 | All | Critical | System can go offline shortly after changing the SMTP settings using the chemailserver command via the GUI (show details) | 8.3.1.6 | System Monitoring |
HU02326 | SVC | High Importance | Delays in passing messages between nodes in an I/O group can adversely impact write performance (show details) | 8.3.1.6 | Performance |
HU02345 | All | High Importance | When connectivity to nodes in a local or remote cluster is lost, inflight IO can become stuck in an aborting state, consuming system resources and potentially adversely impacting performance (show details) | 8.3.1.6 | HyperSwap, Metro Mirror |
HU02362 | FS5100, FS7200, FS9100, FS9200, SVC, V5100, V7000 | High Importance | When the RAID scrub process encounters bad grains, the peak response time for reads and writes can be adversely impacted (show details) | 8.3.1.6 | RAID |
HU02376 | All | High Importance | FlashCopy maps may get stuck at 99% due to inconsistent metadata accounting between nodes (show details) | 8.3.1.6 | FlashCopy |
HU02377 | All | High Importance | A race condition in DRP may stop IO being processed leading to timeouts (show details) | 8.3.1.6 | Data Reduction Pools |
HU02392 | All | High Importance | Validation in the Upload Support Package feature will reject new case number formats in the PMR field (show details) | 8.3.1.6 | Support Data Collection |
HU02422 | All | High Importance | GUI performance can be degraded when displaying large numbers of volumes or other objects (show details) | 8.3.1.6 | Graphical User Interface |
IT36792 | All | High Importance | EasyTier can select a default performance profile for a drive which could cause too much hot data to be moved to lower tiers (show details) | 8.3.1.6 | EasyTier |
IT38015 | All | High Importance | During RAID rebuild or copyback on systems with 16gb or less of memory, cache handling can lead to a deadlock which results in timeouts (show details) | 8.3.1.6 | RAID |
HU02331 | All | Suggested | Due to a threshold issue an error code 3400 may appear too often in the event log (show details) | 8.3.1.6 | Compression |
HU02332 & HU02336 | All | Suggested | When an I/O is received, from a host, with invalid or inconsistent SCSI data but a good checksum it may cause a node warmstart (show details) | 8.3.1.6 | Hosts |
HU02366 | All | Suggested | Slow internal resource reclamation by the RAID component can cause a node warmstart (show details) | 8.3.1.6 | RAID |
HU02375 | All | Suggested | An issue in how the GUI handles volume data can adversely impact its responsiveness (show details) | 8.3.1.6 | Graphical User Interface |
HU02399 | SVC | Suggested | Boot drives may be reported as having invalid state by the GUI, even though they are online (show details) | 8.3.1.6 | Graphical User Interface |
HU02419 | All | Suggested | During creation of a drive FRU id the resulting unique number can contain a space character which can lead to CLI commands, that return this value, presenting it as a truncated string (show details) | 8.3.1.6 | Command Line Interface, Drives |
HU02424 | All | Suggested | Frequent GUI refreshing adversely impacts usability on some screens (show details) | 8.3.1.6 | Graphical User Interface |
HU02425 | All | Suggested | An issue in the handling of internal messages, when the system has a high IO workload to two or more different FlashCopy maps in the same dependency chain, can result in incorrect counters. The node will warmstart to clear this condition. (show details) | 8.3.1.6 | FlashCopy |
HU02186 | FS5100, FS7200, FS9100, FS9200, V5100, V7000 | HIPER | NVMe drive pulls or firmware upgrades may lead to offline pools with the possibility of a small loss of data integrity. For more details refer to this Flash (show details) | 8.3.1.5 | RAID |
HU02360 | All | High Importance | Cloud Callhome may stop working and provide no indication of this in the event log. For more details refer to this Flash (show details) | 8.3.1.5 | System Monitoring |
FS5100, FS7200, FS9100, FS9200, V5100, V7000 | HIPER | This APAR has been reverted at this PTF. This APAR will be re-applied in a future PTF | 8.3.1.4 | RAID | |
HU02261 | All | HIPER | A Data Reduction Pool may be taken offline when metadata is detected to hold an invalid compression flag. For more details refer to this Flash (show details) | 8.3.1.4 | Data Reduction Pools |
HU02313 | FS5100, FS7200, FS9100, FS9200, V5100, V7000 | HIPER | When a FlashCore Module (FCM) fails there is a chance that this can trigger other FCMs in the same control enclosure to also fail. If enough additional drives fail, at the same time, this can take the array offline and cause a loss of access to data. For more details refer to this Flash (show details) | 8.3.1.4 | Drives |
HU02338 | All | HIPER | An issue in the setting up of reverse FlashCopy mappings can cause the background copy to finish prematurely providing an incomplete target image (show details) | 8.3.1.4 | FlashCopy |
HU02340 | All | HIPER | High replication workloads can cause multiple warmstarts with a loss of access at the partner cluster (show details) | 8.3.1.4 | IP Replication |
IT35555 | V5000 | HIPER | Storwize V5030 systems running v8.3.1.3 may experience an offline pool under heavy I/O workloads (show details) | 8.3.1.4 | Drives |
HU02282 | All | Critical | After a code upgrade the config node may exhibit high write response times. In exceptionally rare circumstances an Mdisk group may be taken offline (show details) | 8.3.1.4 | Cache |
HU02314 | FS5100, FS7200, FS9100, FS9200, V5100, V7000 | Critical | Due to a RAID issue when a bad block is detected on a NVMe drive there may be multiple node warmstarts with a possibility of a loss of access to data (show details) | 8.3.1.4 | Drives |
HU02315 | All | Critical | Failover for VMware iSER hosts may pause I/O for more than 120 seconds (show details) | 8.3.1.4 | Hosts |
HU02321 | All | Critical | Where nodes relying on RDMA clustering alone, if a node is removed, warmstarts or goes down for upgrade there may be a delay in internode communication leading to lease expiries (show details) | 8.3.1.4 | iSCSI |
HU02322 | All | Critical | A deadlock condition in the Data Reduction Pool function may cause multiple node warmstarts and a temporary loss of access to data (show details) | 8.3.1.4 | Data Reduction Pools |
HU02323 | All | Critical | Stalled I/O during DRAID expansion can cause node warmstarts and a temporary loss of access to data (show details) | 8.3.1.4 | Distributed RAID |
HU02153 | All | High Importance | Fabric or host issues can cause aborted IOs to block the port throttle queue leading to adverse performance that is cleared by a node warmstart (show details) | 8.3.1.4 | Hosts |
HU02227 | FS7200, FS9100, FS9200, SVC, V5100, V7000 | High Importance | Certain I/O patterns can cause compression hardware to post errors. When those errors exceed a threshold the node can be taken offline (show details) | 8.3.1.4 | Compression |
HU02311 | All | High Importance | An issue in volume copy flushing may lead to higher than expected write cache delays (show details) | 8.3.1.4 | Cache |
HU02317 | All | High Importance | A DRAID expansion can stall shortly after it is initiated (show details) | 8.3.1.4 | Distributed RAID |
HU02095 | All | Suggested | The effective_used_capacity field of lsarray/lsmdisk commands should be empty for RAID arrays which do not contain overprovisioned drives. However, sometimes this field can be zero even though it should be empty. This can cause incorrect provisioned capacity reporting in the GUI (show details) | 8.3.1.4 | Graphical User Interface |
HU02280 | All | Suggested | Spectrum Control or Storage Insights may be unable to collect stats after a Tier 2 recovery or system powerdown (show details) | 8.3.1.4 | System Monitoring |
HU02292 & HU02308 | All | Suggested | The use of maximum replication delay within Global Mirror may occasionally cause a node warmstart (show details) | 8.3.1.4 | Global Mirror |
HU02277 | All | HIPER | RAID parity scrubbing can become stalled causing an accumulation of media errors leading to multiple drive failures with the possibility of data integrity loss. For more details refer to this Flash (show details) | 8.3.1.3 | RAID |
HU02058 | All | Critical | Changing a remote copy relationship from GMCV to MM or GM can result in a Tier 2 recovery (show details) | 8.3.1.3 | Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02162 | All | Critical | When a node warmstart occurs during an upgrade from v8.3.0.0, or earlier, to 8.3.0.1, or later, with dedup enabled it can lead to repeated node warmstarts across the cluster necessitating a Tier 3 recovery (show details) | 8.3.1.3 | Data Reduction Pools |
HU02180 | All | Critical | When a svctask restorefcmap command is run on a VVol that is the target of another FlashCopy mapping both nodes in an I/O group may warmstart (show details) | 8.3.1.3 | vVols |
HU02184 | All | Critical | When a 3PAR controller experiences a fault that prevents normal I/O processing it may issue a SCSI TARGET RESET command. This command is not supported and may cause multiple node asserts, possibly cluster-wide (show details) | 8.3.1.3 | Backend Storage |
HU02196 & HU02253 | All | Critical | A particular sequence of internode messaging delays can lead to a cluster wide lease expiry (show details) | 8.3.1.3 | Reliability Availability Serviceability |
HU02210 | All | Critical | There is a very small timing window where a volume may be reported as offline, to a host, during its conversion from a regular volume to a HyperSwap volume (show details) | 8.3.1.3 | HyperSwap |
HU02213 | SVC | Critical | A Hot Spare Node (HSN) timing window issue can, during an HSN activation or deactivation, cause the cluster to broadcast an invalid VPD update to other clusters on the SAN. This may trigger a Tier 2 recovery on the other cluster. For more details refer to this Flash (show details) | 8.3.1.3 | Hot Spare Node |
HU02230 | FS7200, FS9100, FS9200, V7000 | Critical | For IBM Flash Core Modules a change of state, from unused to candidate, can lead to a Tier 2 recovery (show details) | 8.3.1.3 | Drives |
HU02262 | SVC, V5000, V7000 | Critical | Entering the CLI applydrivesoftware -cancel command may result in cluster-wide warmstarts (show details) | 8.3.1.3 | Drives |
HU02266 | All | Critical | An issue in auto-expand can cause expansion to fail and the volume to be taken offline (show details) | 8.3.1.3 | Thin Provisioning |
HU02289 | FS9200, SVC | Critical | An issue with internal resource allocation in high-end systems, with 1000s of mirror copies, may cause multiple warmstarts with the possibility of a loss of access (show details) | 8.3.1.3 | Volume Mirroring |
HU02295 | SVC | Critical | When upgrading from v8.2.1 or v8.3, in the presence of hot spare nodes, an issue with the handling of node metadata may cause a Tier 2 recovery (show details) | 8.3.1.3 | System Update |
HU02390 | All | Critical | A memory handling issue in the REST API may cause an out-of-memory condition when listing a large number of volumes (show details) | 8.3.1.3 | REST API |
HU02156 | All | High Importance | Global Mirror environments may experience more frequent 1920 events due to writedone message queuing (show details) | 8.3.1.3 | Global Mirror |
HU02164 | All | High Importance | An issue in Remote Copy may cause a loss of hardened data when a node is warmstarted (show details) | 8.3.1.3 | Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU02194 | All | High Importance | Password reset via USB drive does not work as expected and user is not able to login to Management or Service assistant GUI with the new password (show details) | 8.3.1.3 | Reliability Availability Serviceability |
HU02201 & HU02221 | All | High Importance | Shortly after upgrading drive firmware, specific drive models can fail due to Too many long IOs to drive for too long errors (show details) | 8.3.1.3 | Drives |
HU02248 | All | High Importance | After upgrade the system may be unable to perform LDAP authentication (show details) | 8.3.1.3 | LDAP |
II14767 | SVC | High Importance | An issue with how cache handles ownership of volumes across multiple sites can lead to cross-site destage, adversely impacting write latency. For more details refer to this Flash (show details) | 8.3.1.3 | Cache |
HU02142 | All | Suggested | It is possible for a backend unmap process to become stalled, preventing system configuration changes from completing (show details) | 8.3.1.3 | Distributed RAID |
HU02208 | All | Suggested | An issue with the handling of files by quorum can lead to a node warmstart (show details) | 8.3.1.3 | Quorum |
HU02241 | All | Suggested | IP Replication can fail to create IP partnerships via the secondary cluster management IP (show details) | 8.3.1.3 | IP Replication |
HU02244 | SVC | Suggested | False positive node error 766 (depleted CMOS battery) messages may appear in the Event Log (show details) | 8.3.1.3 | System Monitoring |
HU02251 | All | Suggested | A warmstart may occur when a node receives iSCSI host login/logout requests out of sequence (show details) | 8.3.1.3 | Hosts, iSCSI |
HU02255 | All | Suggested | A timing issue in the processing of login requests can cause a single node warmstart (show details) | 8.3.1.3 | Command Line Interface, Graphical User Interface |
HU02281 | All | Suggested | When upgrading from v8.2.1, or earlier, to v8.3.0, or later, the CLI and GUI may incorrectly show all hosts offline. Checks from the host perspective will show them to be online (show details) | 8.3.1.3 | Hosts, System Update |
HU02303 & HU02305 | All | Suggested | Configuration node warmstart will occur if mkhostcluster is run with -ignoreseedvolume and the ignored volumes have an id greater than 256 (show details) | 8.3.1.3 | Hosts |
HU02358 | All | Suggested | An issue in Remote Copy, that stalls a switch of direction, can cause I/O timeouts leading to a node warmstart (show details) | 8.3.1.3 | Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
IT32338 | All | Suggested | Testing LDAP Authentication fails if username & password are supplied (show details) | 8.3.1.3 | LDAP |
HU02182 | All | HIPER | Cisco MDS switches with old firmware may refuse port logins leading to a loss of access. For more details refer to this Flash (show details) | 8.3.1.2 | Backend Storage, Hosts, System Update |
HU02186 (reverted in 8.3.1.4) | FS5100, FS7200, FS9100, FS9200, V5100, V7000 | HIPER | NVMe drive pulls or firmware upgrades may lead to offline pools with the possibility of a small loss of data integrity. For more details refer to this Flash (show details) | 8.3.1.2 | RAID |
HU02212 | All | HIPER | Remote Copy secondary may have inconsistent data following a stop with -access due to a missing bitmap merge from FlashCopy to Remote Copy. For more details refer to this Flash (show details) | 8.3.1.2 | Global Mirror With Change Volumes, HyperSwap |
HU02234 | All | HIPER | An issue in HyperSwap Read Passthrough can cause multiple node warmstarts with the possibility of a loss of access to data (show details) | 8.3.1.2 | HyperSwap |
HU02237 | All | HIPER | Under a rare and complicated set of conditions, a RAID 1 or RAID 10 array may drop a write, causing undetected data corruption. For more details refer to this Flash (show details) | 8.3.1.2 | RAID |
HU02238 | All | HIPER | Force-stopping a FlashCopy map, where the source volume is a Metro or Global Mirror target volume, may cause other FlashCopy maps to return invalid data if they are not 100% copied, in specific configurations. For more details refer to this Flash (show details) | 8.3.1.2 | FlashCopy, Global Mirror, Metro Mirror |
HU01968 & HU02215 | All | Critical | An upgrade may fail due to corrupt hardened data in a node. This can affect an I/O group (show details) | 8.3.1.2 | System Update |
HU02106 | All | Critical | Multiple node warmstarts, in quick succession, can cause the partner node to lease expire (show details) | 8.3.1.2 | IP Quorum, Quorum |
HU02135 | All | Critical | Removing multiple IQNs for an iSCSI host can result in a Tier 2 recovery (show details) | 8.3.1.2 | iSCSI |
HU02154 | All | Critical | If a node is rebooted, when remote support is enabled, then all other nodes will warmstart (show details) | 8.3.1.2 | Support Remote Assist |
HU02202 | All | Critical | During an migratevdisk operation if MDisk tiers in the target pool do not match those in the source pool then a Tier 2 recovery may occur (show details) | 8.3.1.2 | EasyTier |
HU02207 | All | Critical | If hosts send more concurrent iSCSI commands than a node can handle then it may enter a service state (error 578) (show details) | 8.3.1.2 | iSCSI |
HU02216 | All | Critical | When migrating or deleting a Change Volume of a RC relationship the system might be exposed to a Tier 2 (Automatic Cluster Restart) recovery. When deleting the Change Volumes, the T2 will re-occur which will place the nodes into a 564 state. The migration of the Change Volume will trigger a T2 and recover. For more details refer to this Flash (show details) | 8.3.1.2 | Global Mirror With Change Volumes |
HU02222 | All | Critical | Where the source volume of an incremental FlashCopy map is also a Metro or Global Mirror target volume that is using a change volume or is a Hyperswap volume, then there is a possibility that not all data will be copied to the FlashCopy target. For more details refer to this Flash (show details) | 8.3.1.2 | Global Mirror With Change Volumes |
HU02242 | All | Critical | An iSCSI IP address, with a gateway argument of 0.0.0.0, is not properly assigned to each Ethernet port and any previously set iSCSI IP address may be retained (show details) | 8.3.1.2 | iSCSI |
IT32631 | All | Critical | Whilst upgrading the firmware for multiple drives an issue in the firmware checking can initiate a Tier 2 recovery (show details) | 8.3.1.2 | Drives |
HU02128 | All | High Importance | Deduplication volume lookup can over utilise resources causing an adverse performance impact (show details) | 8.3.1.2 | Data Reduction Pools, Deduplication |
HU02168 | V5000, V7000 | High Importance | In the event of unexpected power loss a node may not save system data (show details) | 8.3.1.2 | Reliability Availability Serviceability |
HU02203 | FS9100, V5000, V7000 | High Importance | When a node reboots, it is possible for the node to be unable to communicate with some of the NVMe drives in the enclosure (show details) | 8.3.1.2 | Drives |
HU02204 | All | High Importance | After a Tier 2 recovery a node may fail to rejoin the cluster (show details) | 8.3.1.2 | Reliability Availability Serviceability |
HU02229 | FS5000, V5000 | High Importance | An issue in the BIOS firmware of some systems can cause a severe performance impact for iSCSI hosts (show details) | 8.3.1.2 | iSCSI |
HU01931 | SVC, V7000 | Suggested | Where a high rate of CLI commands are received, it is possible for inter-node processing code to be delayed which results in a small increase in receive queue time on the config node (show details) | 8.3.1.2 | Performance |
HU02015 | FS9100, V5000, V7000 | Suggested | Some read-intensive SSDs are incorrectly reporting wear rate thresholds generating unnecessary errors in the Event Log (show details) | 8.3.1.2 | Drives |
HU02091 | V5000 | Suggested | Upgrading to v8.2.1.8, or later, may result in a licensing error in the Event Log (show details) | 8.3.1.2 | Licensing |
HU02137 | All | Suggested | An issue with support for target resets in Nimble Storage controllers may cause a node warmstart (show details) | 8.3.1.2 | Backend Storage |
HU02175 | All | Suggested | A GUI issue can cause drive counts to be inconsistent and crash browsers (show details) | 8.3.1.2 | Graphical User Interface |
HU02178 | All | Suggested | IP Quorum hosts may not be shown in lsquorum command output (show details) | 8.3.1.2 | IP Quorum |
HU02224 | All | Suggested | When the RAID component fails to free up memory quickly enough for I/O processing there can be a single node warmstart (show details) | 8.3.1.2 | RAID |
HU02235 | All | Suggested | The SSH CLI prompt can contain the characters FB after the cluster name (show details) | 8.3.1.2 | Command Line Interface |
HU02341 | All | Suggested | Cloud Callhome can become disabled due to an internal issue. A related error may not being recorded in the event log (show details) | 8.3.1.2 | System Monitoring |
IT32440 | All | Suggested | Under heavy I/O workload the processing of deduplicated I/O may cause a single node warmstart (show details) | 8.3.1.2 | Deduplication |
IT32519 | All | Suggested | Changing an LDAP users password, in the directory, whilst this user is logged in to the GUI of a Spectrum Virtualize system may result in an account lockout in the directory, depending on the account lockout policy configured for the directory. Existing CLI logins via SSH are not affected (show details) | 8.3.1.2 | LDAP |
HU01894 | All | HIPER | After node reboot, or warmstart, some volumes accessed by AIX, VIO or VMware hosts may experience stuck SCSI2 reservations on the NPIV failover ports of the partner node. This can cause a loss of access to data (show details) | 8.3.1.0 | Hosts |
HU02075 | All | HIPER | A FlashCopy snapshot, sourced from the target of an Incremental FlashCopy map, can sometimes, temporarily, present incorrect data to the host (show details) | 8.3.1.0 | FlashCopy |
HU02141 | All | HIPER | An issue in the max replication delay function may trigger a Tier 2 recovery, after posting multiple 1920 errors in the Event Log. For more details refer to this Flash (show details) | 8.3.1.0 | Global Mirror |
HU02205 | All | HIPER | Incremental FlashCopy targets can be corrupted when the FlashCopy source is a target of a remote copy relationship (show details) | 8.3.1.0 | FlashCopy, Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU01967 | All | Critical | When I/O, in remote copy relationships, experiences delays (1720 and/or 1920 errors are logged) an I/O group may warmstart (show details) | 8.3.1.0 | Global Mirror, Global Mirror With Change Volumes, Metro Mirror |
HU01970 | All | Critical | When a GMCV relationship is stopped, with the -access option, and the secondary volume is immediately deleted with -force, then all nodes may repeatedly warmstart (show details) | 8.3.1.0 | Global Mirror With Change Volumes |
HU02017 | All | Critical | Unstable inter-site links may cause a system-wide lease expiry leaving all nodes in a service state - one with error 564 and others with error 551 (show details) | 8.3.1.0 | Reliability Availability Serviceability |
HU02054 | All | Critical | The event log handler maintains a second list of events. On rare occasions, for log full events, these lists can get out of step, resulting in a Tier 2 recovery (show details) | 8.3.1.0 | System Monitoring |
HU02063 | All | Critical | HyperSwap clusters with only two surviving nodes may experience warmstarts on both of those nodes where rcbuffersize is set to 512MB (show details) | 8.3.1.0 | HyperSwap |
HU02065 | All | Critical | Mishandling of Data Reduction Pool allocation request rejections can lead to node warmstarts that can take an MDisk group offline (show details) | 8.3.1.0 | Data Reduction Pools |
HU02066 | All | Critical | If, during large (>8KB) reads from a host, a medium error is encountered, on backend storage, then there may be node warmstarts, with the possibility of a loss of access to data (show details) | 8.3.1.0 | Data Reduction Pools |
HU02108 | All | Critical | Deleting a managed disk group, with -force, may cause multiple warmstarts with the possibility of a loss of access to data (show details) | 8.3.1.0 | Data Reduction Pools |
HU02109 | All | Critical | Free extents may not be unmapped after volume deletion, or migration, resulting in out-of-space conditions on backend controllers (show details) | 8.3.1.0 | Backend Storage, SCSI Unmap |
HU02115 | All | Critical | Attempting to upgrade all drive firmware, with an inadequate drive package, may lead to multiple node warmstarts, with the possibility of a loss of access to data (show details) | 8.3.1.0 | Drives |
HU02138 | All | Critical | An issue in Data Reduction Pool garbage collection can cause I/O timeouts leading to an offline pool (show details) | 8.3.1.0 | Data Reduction Pools |
HU02152 | All | Critical | Due to an issue in RAID there may be I/O timeouts, leading to node warmstarts, with the possibility of a loss of access to data (show details) | 8.3.1.0 | RAID |
HU02197 | All | Critical | Bulk volume removals can adversely impact related FlashCopy mappings leading to a Tier 2 recovery (show details) | 8.3.1.0 | FlashCopy |
IT29867 | All | Critical | If a change volume, for a remote copy relationship, in a consistency group, runs out of space whilst properties, of the consistency group, are being changed then a Tier 2 recovery may occur (show details) | 8.3.1.0 | Global Mirror With Change Volumes |
IT31113 | All | Critical | After a manual power off and on, of a system, both nodes, in an I/O group, may repeatedly assert into a service state (show details) | 8.3.1.0 | RAID |
IT31300 | All | Critical | When a snap collection reads the status of PCI devices a CPU can be stalled leading to a cluster-wide lease expiry (show details) | 8.3.1.0 | Support Data Collection |
HU01890 | All | High Importance | FlashCopy mappings, from master volume to primary change volume, may become stalled when a T2 recovery occurs whilst the mappings are in a copying state (show details) | 8.3.1.0 | Global Mirror With Change Volumes |
HU01923 | All | High Importance | An issue in the way Global Mirror handles write sequence numbers >512 may cause multiple node warmstarts (show details) | 8.3.1.0 | Global Mirror |
HU01964 | All | High Importance | An issue in the cache component may limit I/O throughput (show details) | 8.3.1.0 | Cache |
HU02037 | All | High Importance | A FlashCopy consistency group, with a mix of mappings in different states, cannot be stopped (show details) | 8.3.1.0 | FlashCopy |
HU02078 | SVC | High Importance | Heavily unbalanced workloads, in stretched-cluster configurations, can bias inter-node traffic through one port, adversely affecting performance (show details) | 8.3.1.0 | Performance |
HU02114 | FS5000, FS7200, FS9100, FS9200, V7000 | High Importance | Upgrading FCM firmware on multiple I/O group systems can cause a drive to become stuck at 0% sync with the corresponding array in a 'syncing' state (show details) | 8.3.1.0 | Drives |
HU02132 | All | High Importance | Removing a thin-provisioned volume and then immediately creating one of the same size may cause node warmstarts (show details) | 8.3.1.0 | Thin Provisioning |
HU02143 | All | High Importance | The performance profile, for some enterprise tier drives, may not correctly match the drives capabilities leading to that tier being overdriven (show details) | 8.3.1.0 | EasyTier |
HU02169 | All | High Importance | After a Tier 3 recovery, different nodes may report different UIDs for a subset of volumes (show details) | 8.3.1.0 | Hosts |
HU02206 | All | High Importance | Garbage collection can operate at inappropriate times, generating inefficient backend workload, adversely affecting flash drive write endurance and overloading nearline drives (show details) | 8.3.1.0 | Data Reduction Pools |
HU01746 | All | Suggested | Adding a volume copy may deactivate any associated MDisk throttling (show details) | 8.3.1.0 | Throttling |
HU01796 | All | Suggested | Battery Status LED may not illuminate (show details) | 8.3.1.0 | System Monitoring |
HU01891 | All | Suggested | An issue in DRAID grain process scheduling can lead to a duplicate entry condition that is cleared by a node warmstart (show details) | 8.3.1.0 | Distributed RAID |
HU01943 | All | Suggested | Stopping a GMCV relationship with the -access flag may result in more processing than is required (show details) | 8.3.1.0 | Global Mirror With Change Volumes |
HU01953 | All | Suggested | Following a Data Reduction Pool recovery, in some circumstances, it may not be possible to create new volumes, via the GUI, due to an incorrect value being returned from the lsmdiskgrp (show details) | 8.3.1.0 | Graphical User Interface |
HU02021 | All | Suggested | Disabling garbage collection may cause a node warmstart (show details) | 8.3.1.0 | Data Reduction Pools |
HU02023 | All | Suggested | An issue with the processing of FlashCopy map commands may result in a single node warmstart (show details) | 8.3.1.0 | Command Line Interface, FlashCopy |
HU02026 | All | Suggested | A timing window issue in the processing of FlashCopy status listing commands can cause a node warmstart (show details) | 8.3.1.0 | Command Line Interface, FlashCopy |
HU02040 | V5000 | Suggested | VPD contains the incorrect FRU part number for the SAS adapter (show details) | 8.3.1.0 | Reliability Availability Serviceability |
HU02048 | All | Suggested | An issue in the handling of ATS commands from VMware hosts can cause a single node warmstart (show details) | 8.3.1.0 | Hosts |
HU02052 | All | Suggested | During an upgrade an issue, with buffer handling, in Data Reduction Pool can lead to a node warmstart (show details) | 8.3.1.0 | Data Reduction Pools |
HU02062 | All | Suggested | An issue, with node index numbers for I/O groups, when using 32Gb HBAs may result in host ports incorrectly being reported offline (show details) | 8.3.1.0 | Hosts |
HU02085 | All | Suggested | Freeze time of Global Mirror remote copy consistency groups may not be updated correctly in certain scenarios (show details) | 8.3.1.0 | Global Mirror |
HU02099 | All | Suggested | Cloud callhome error 3201 messages may appear in the Event Log (show details) | 8.3.1.0 | System Monitoring |
HU02102 | All | Suggested | Excessive processing time required for FlashCopy bitmap operations, associated with large (> 20TB) Global Mirror change volumes, may lead to a node warmstart (show details) | 8.3.1.0 | Global Mirror With Change Volumes |
HU02103 | FS9100, V5000, V7000 | Suggested | The system management firmware may, incorrectly, attempt to obtain an IP address, using DHCP, making it accessible via Ethernet (show details) | 8.3.1.0 | |
HU02111 | All | Suggested | An issue with how Data Reduction Pool handles data, at the sub-extent level, may result in a node warmstart (show details) | 8.3.1.0 | Data Reduction Pools |
HU02119 | FS5000, FS7200, FS9100, FS9200, V7000 | Suggested | NVMe drive replacement on 8.3.0.0 or 8.3.0.1 may result in the GUI, and lsdrive CLI command, showing a ghost drive (show details) | 8.3.1.0 | Drives |
HU02146 | All | Suggested | An issue in inter-node message handling may cause a node warmstart (show details) | 8.3.1.0 | Reliability Availability Serviceability |
HU02157 | All | Suggested | Issuing a mkdistributedarray command may result in a node warmstart (show details) | 8.3.1.0 | Distributed RAID |
HU02173 | All | Suggested | During a pending fabric login, when an abort is received, it is possible for a related entry in the WWPN table to not be removed. The node will warmstart to clear this condition (show details) | 8.3.1.0 | Reliability Availability Serviceability |
HU02183 | All | Suggested | An issue in the way inter-node communication is handled can lead to a node warmstart (show details) | 8.3.1.0 | Reliability Availability Serviceability |
HU02190 | All | Suggested | Error 1046 not triggering a Call Home even though it is a hardware fault (show details) | 8.3.1.0 | System Monitoring |
HU02214 | All | Suggested | Under a certain I/O pattern it is possible for metadata management in Data Reduction Pools to become inconsistent leading to a node warmstart (show details) | 8.3.1.0 | Data Reduction Pools |
HU02247 | All | Suggested | Unnecessary Ethernet MAC flapping messages reported in switch logs (show details) | 8.3.1.0 | Reliability Availability Serviceability |
HU02285 | All | Suggested | Single node warmstart due to cache resource allocation issue (show details) | 8.3.1.0 | Cache |
IT21896 | All | Suggested | Where encryption keys have been lost it will not be possible to remove an empty MDisk group (show details) | 8.3.1.0 | Encryption |
IT30306 | All | Suggested | A timing issue in callhome function initialisation may cause a node warmstart (show details) | 8.3.1.0 | System Monitoring |
Description | Link |
---|---|
Support Websites | |
Update Matrices, including detailed build version | |
Support Information pages providing links to the following information:
|
|
Spectrum Virtualize Family of Products Inter-System Metro Mirror and Global Mirror Compatibility Cross Reference | |
Software Upgrade Test Utility | |
Software Upgrade Planning |