ESXi 6.5 Update 2 | 3 MAY 2018 | ISO Build 8294253 Check for additions and updates to these release notes. |
What's in the Release Notes
The release notes cover the following topics:
- What's New
- Earlier Releases of ESXi 6.5
- Internationalization
- Compatibility
- Product and Support Notices
- Patches Contained in this Release
- Resolved Issues
- Known Issues
What's New
The ESXi 6.5 Update 2 release includes the following list of new features.
- ESXi 6.5 Update 2 enables Microsemi Smart PQI (smartpqi) driver support for the HPE ProLiant Gen10 Smart Array Controller.
- Updates to the nhpsa driver: The disk serviceability plug-in of the ESXi native driver for HPE Smart Array controllers, nhpsa, now works with an extended list of expander devices to enable compatibility with HPE Gen9 configurations.
- Updates to the nvme driver: ESXi 6.5 Update 2 adds management of multiple namespaces compatible with the Non-Volatile Memory Express (NVMe) 1.2 specification and enhanced diagnostic log.
- With ESXi 6.5 Update 2, you can add tags to the Trusted Platform Module (TPM) hardware version 1.2 on ESXi by using ESXCLI commands. A new API provides read and write access in the form of
get
,set
, andclear
commands to the non-volatile memory on TPMs of version 1.2. - ESXi 6.5 Update 2 enables a new native driver to support the Broadcom SAS 3.5 IT/IR controllers with devices including combination of NVMe, SAS, and SATA drives. The HBA device IDs are: SAS3516(00AA), SAS3516_1(00AB), SAS3416(00AC), SAS3508(00AD), SAS3508_1(00AE), SAS3408(00AF), SAS3716(00D0), SAS3616(00D1), SAS3708(00D2).
- In ESXi 6.5 Update 2, LightPulse and OneConnect adapters are supported by separate default drivers. The brcmfcoe driver supports OneConnect adapters and the lpfc driver supports only LightPulse adapters. Previously, the lpfc driver supported both OneConnect and LightPulse adapters.
Earlier Releases of ESXi 6.5
Features and known issues of ESXi 6.5 are described in the release notes for each release. Release notes for earlier releases of ESXi 6.5 are:
- VMware ESXi 6.5 Update 1 Release Notes
- VMware ESXi 6.5.0a Release Notes
- VMware vSphere 6.5 Release Notes
For compatibility, installation and upgrades, product support notices, and features see the VMware vSphere 6.5 Release Notes.
Internationalization
VMware vSphere 6.5 is available in the following languages:
- English
- French
- German
- Spanish
- Japanese
- Korean
- Simplified Chinese
- Traditional Chinese
Components of VMware vSphere 6.5, including vCenter Server, ESXi, the vSphere Web Client, and the vSphere Client do not accept non-ASCII input.
Compatibility
ESXi, vCenter Server, and vSphere Web Client Version Compatibility
The VMware Product Interoperability Matrix provides details about the compatibility of current and earlier versions of VMware vSphere components, including ESXi, VMware vCenter Server, the vSphere Web Client, and optional VMware products. Check the VMware Product Interoperability Matrix also for information about supported management and backup agents before you install ESXi or vCenter Server.
The vSphere Web Client and vSphere Client are packaged with vCenter Server.
Hardware Compatibility for ESXi
To view a list of processors, storage devices, SAN arrays, and I/O devices that are compatible with vSphere 6.5 Update 2, use the ESXi 6.5 information in the VMware Compatibility Guide.
Device Compatibility for ESXi
To determine which devices are compatible with ESXi 6.5, use the ESXi 6.5 information in the VMware Compatibility Guide.
Guest Operating System Compatibility for ESXi
To determine which guest operating systems are compatible with vSphere 6.5, use the ESXi 6.5 information in the VMware Compatibility Guide.
Virtual Machine Compatibility for ESXi
Virtual machines that are compatible with ESX 3.x and later (hardware version 4) are supported with ESXi 6.5. Virtual machines that are compatible with ESX 2.x and later (hardware version 3) are not supported. To use such virtual machines on ESXi 6.5, upgrade the virtual machine compatibility.
See the vSphere Upgrade documentation.
Product and Support Notices
Important: Upgrade paths from ESXi 6.5 Update 2 to ESXi 6.7 are not supported.
- The VMware Lifecycle Product Matrix provides detailed information about all supported and unsupported products. Check the VMware Lifecycle Product Matrix also for further information about the End of General Support, End of Technical Guidance, and End Of Availability.
Patches Contained in this Release
This release contains all bulletins for ESXi that were released earlier to the release date of this product. See the My VMware page for more information about the individual bulletins.
Details
Download Filename: | update-from-esxi6.5-6.5_update02.zip |
Build: | 8294253 8285314 (Security-only) |
Download Size: | 486.6 MB |
md5sum: |
45a251b8c8d80289f678041d5f1744a3 |
sha1checksum: |
fa4aa02e4bf06babca214be2dcd61facbd42c493 |
Host Reboot Required: | Yes |
Virtual Machine Migration or Shutdown Required: | Yes |
For more information, see VMware ESXi 6.5, Patch Release ESXi-6.5.0-update02.
Bulletins
This release contains general and security-only bulletins. Security-only bulletins are applicable to new security fixes only. No new bug fixes are included, but bug fixes from earlier patch and update releases are included.
If the installation of all new security and bug fixes is required, you must apply all bulletins in this release. In some cases, the general release bulletin will supersede the security- only bulletin. This is not an issue as the general release bulletin contains both the new security and bug fixes.
The security-only bulletins are identified by bulletin IDs that end in "SG". For information on patch and update classification, see KB 2014447.
Bulletin ID | Category | Severity | Knowledge Base Article |
ESXi650-201805201-UG | Bugfix | Critical | 53067 |
ESXi650-201805202-UG | Bugfix | Important | 53068 |
ESXi650-201805203-UG | Bugfix | Important | 53069 |
ESXi650-201805204-UG | Bugfix | Important | 53070 |
ESXi650-201805205-UG | Enhancement | Important | 53071 |
ESXi650-201805206-UG | Enhancement | Important | 53072 |
ESXi650-201805207-UG | Enhancement | Important | 53073 |
ESXi650-201805208-UG | Enhancement | Important | 53074 |
ESXi650-201805209-UG | Enhancement | Important | 53075 |
ESXi650-201805210-UG | Enhancement | Important | 53076 |
ESXi650-201805211-UG | Enhancement | Important | 53077 |
ESXi650-201805212-UG | Bugfix | Important | 53078 |
ESXi650-201805213-UG | Enhancement | Important | 53079 |
ESXi650-201805214-UG | Bugfix | Important | 53080 |
ESXi650-201805215-UG | Bugfix | Important | 53081 |
ESXi650-201805216-UG | Bugfix | Important | 53082 |
ESXi650-201805217-UG | Bugfix | Important | 53083 |
ESXi650-201805218-UG | Enhancement | Important | 53084 |
ESXi650-201805219-UG | Bugfix | Important | 53085 |
ESXi650-201805220-UG | Bugfix | Important | 53086 |
ESXi650-201805221-UG | Bugfix | Moderate | 53087 |
ESXi650-201805222-UG | Bugfix | Moderate | 53088 |
ESXi650-201805223-UG | Bugfix | Important | 53089 |
ESXi650-201805101-SG | Security | Important | 53092 |
ESXi650-201805102-SG | Security | Important | 53093 |
ESXi650-201805103-SG | Security | Important | 53094 |
Image Profiles
VMware patch and update releases contain general and critical image profiles.
Application of the general release image profile applies to new bug fixes.
Image Profile Name | Knowledge Base Article |
ESXi-6.5.0-20180502001-standard | 53090 |
ESXi-6.5.0-20180502001-no-tools | 53091 |
ESXi-6.5.0-20180501001s-standard | 53095 |
ESXi-6.5.0-20180501001s-no-tools | 53096 |
Resolved Issues
The resolved issues are grouped as follows.
- ESXi650-201805201-UG
- ESXi650-201805202-UG
- ESXi650-201805203-UG
- ESXi650-201805204-UG
- ESXi650-201805205-UG
- ESXi650-201805206-UG
- ESXi650-201805207-UG
- ESXi650-201805208-UG
- ESXi650-201805209-UG
- ESXi650-201805210-UG
- ESXi650-201805211-UG
- ESXi650-201805212-UG
- ESXi650-201805213-UG
- ESXi650-201805214-UG
- ESXi650-201805215-UG
- ESXi650-201805216-UG
- ESXi650-201805217-UG
- ESXi650-201805218-UG
- ESXi650-201805219-UG
- ESXi650-201805220-UG
- ESXi650-201805221-UG
- ESXi650-201805222-UG
- ESXi650-201805223-UG
- ESXi650-201805101-SG
- ESXi650-201805102-SG
- ESXi650-201805103-SG
- ESXi-6.5.0-20180502001-standard
- ESXi-6.5.0-20180502001-no-tools
- ESXi-6.5.0-20180501001s-standard
- ESXi-6.5.0-20180501001s-no-tools
Patch Category | Bugfix |
Patch Severity | Critical |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 1882261, 1907025, 1911721, 1913198, 1913483, 1919300, 1923173, 1937646, 1947334, 1950336, 1950800, 1954364, 1958600, 1962224, 1962739, 1971168, 1972487, 1975787, 1977715, 1982291, 1983123, 1986020, 1987996, 1988248, 1995640, 1996632, 1997995, 1998437, 1999439, 2004751, 2006929, 2007700, 2008663, 2011392, 2012602, 2015132, 2016411, 2016413, 2018046, 2020615, 2020785, 2021819, 2023740, 2023766, 2023858, 2024647, 2028736, 2029476, 2032048, 2033471, 2034452, 2034625, 2034772, 2034800, 2034930, 2039373, 2039896, 2046087, 2048702, 2051775, 2052300, 2053426, 2054940, 2055746, 2057189, 2057918, 2058697, 2059300, 2060589, 2061175, 2064616, 2065000, 1491595, 1994373 |
numbers | N/A |
This patch updates esx-base, esx-tboot,vsan and vsanhealth VIBs to resolve the following issues:
- PR 1913483: Third-party storage array user interfaces might not properly display relationships between virtual machine disks and virtual machines
Third-party storage array user interfaces might not properly display relationships between virtual machine disks and virtual machines when one disk is attached to many virtual machines, due to insufficient metadata.
This issue is resolved in this release.
- PR 1975787: An ESXi support bundle generated by using the VMware Host Client might not contain all log files
Sometimes, when you generate a support bundle for an ESXi host by using the Host Client, the support bundle might not contain all log files.
This issue is resolved in this release.
- PR 1947334: Attempts to modify the vSAN iSCSI parameter VSAN-iSCSI.ctlWorkerThreads fail
You cannot modify the vSAN iSCSI parameter
VSAN-iSCSI.ctlWorkerThreads
, because the parameter is designed as an internal parameter and is not a subject to change. This fix hides internal parameters of vSAN iSCSI to avoid confusion.This issue is resolved in this release.
- PR 1919300: Host Profile compliance check might show Unknown status if a profile is imported from a different vCenter Server system
If you export a Host Profile from one vCenter Server system and import it to another vCenter Server system, compliance checks might show a status Unknown.
This issue is resolved in this release.
- PR 2007700: The advanced setting VMFS.UnresolvedVolumeLiveCheck might be inactive in the vSphere Web Client
In the vSphere Web Client, the option to enable or disable checks during unresolved volume queries with the VMFS.UnresolvedVolumeLiveCheck option under Advance Setting > VMFS, might be grayed out.
This issue is resolved in this release. You can also workaround it by using the vSphere Client or set the option by directly connecting to the ESXi host.
- PR 1983123: ESXi hosts with scratch partitions located on a vSAN datastore might intermittently become unresponsive
If the scratch partition of an ESXi host is located on a vSAN datastore, the host might become unresponsive, because if the datastore is temporary inaccessible, this might block userworld daemons, such as hostd, if they try to access this datastore at the same time. With this fix, scratch partitions to a vSAN datastore are no longer configurable through the API or the user interface.
This issue is resolved in this release.
- PR 1950336: Hostd might stop responding in attempt to reconfigure a virtual machine
Hostd might stop responding in attempt to reconfigure virtual machines. Configuration entries with unset value in the ConfigSpec object, which encapsulates configuration settings for creation and configuration of virtual machines, cause the issue.
This issue is resolved in this release.
- PR 2006929: VMware vSphere vMotion might fail due to timeout and inaccessible virtual machines
vSphere vMotion might fail due to timeout, because virtual machines might need more buffer to store the shared memory page details used by 3D or graphic devices. The issue is observed mainly on virtual machines running graphics intensive loads or working in virtual desktop infrastructure environments. With this fix, the buffer size is increased by four times for 3D virtual machines and if the increased capacity is still not enough, the virtual machine is powered off.
This issue is resolved in this release.
- PR 1958600: The netdump IP address might be lost after upgrading from ESXi 5.5 Update 3b to ESXi 6.0 Update 3 and/or ESXi 6.5 Update 2
The netdump IP address might be lost after your upgrade from ESXi 5.5 Update 3b to ESXi 6.0 Update 3 and/or ESXi 6.5 Update 2.
This issue is resolved in this release.
- PR 1971168: ESXi system identification information in the vSphere Web Client might not be available
System identification information consists of asset tags, service tags, and OEM strings. In earlier releases, this information comes from the Common Information Model (CIM) service. Starting with vSphere 6.5, the CIM service is turned off by default, unless any third-party CIM providers are installed, which prevents ESXi system identification information to display in the vSphere Web Client.
This issue is resolved in this release.
- PR 2020785: Virtual machines created by the VMware Integrated OpenStack might not work properly
When you create virtual machine disks with the VMware Integrated OpenStack block storage service (Cinder), the disks get random UUIDs. In around 0.5% of the cases, the virtual machines do not recognize these UUIDs, which results in unexpected behavior of the virtual machines.
This issue is resolved in this release.
- PR 2004751: A virtual machine might fail to boot after an upgrade to macOS 10.13
When you upgrade a virtual machine to macOS 10.13, the virtual machine might fail to locate a bootable operating system. If the macOS 10.13 installer determines that the system volume is located on a Solid State Disk (SSD), during the OS upgrade, the installer converts the Hierarchical File System Extended (HFS+/HFSX) to the new Apple File System (APFS) format. As a result, the virtual machine might fail to boot.
This issue is resolved in this release.
- PR 2015132: You might not be able to apply a Host Profile
You might not be able to apply a host profile, if the host profile contains a DNS configuration for a VMkernel network adapter that is associated to a vSphere Distributed Switch.
This issue is resolved in this release.
- PR 2011392: An ESXi host might fail with a purple diagnostic screen due to a deadlock in vSphere Distributed Switch health checks
If you enable VLAN and maximum transmission unit (MTU) checks in vSphere Distributed Switches, your ESXi hosts might fail with a purple diagnostic screen due to a possible deadlock.
This issue is resolved in this release.
- PR 2004948: Update to the nhpsa driver
The nhpsa driver is updated to 2.0.22-1vmw.
- PR 1907025: An ESXi host might fail with an error Unexpected message type
An ESXi host might fail with a purple diagnostic screen due to invalid network input in the error message handling of vSphere Fault Tolerance, which results into an
Unexpected message type
error.This issue is resolved in this release.
- PR 1962224: Virtual machines might fail with an error Unexpected FT failover due to PANIC
When you run vSphere Fault Tolerance in vSphere 6.5 and vSphere 6.0, virtual machines might fail with an error such as Unexpected FT failover due to PANIC error: VERIFY bora/vmcore/vmx/main/monitorAction.c:598. The error occurs when virtual machines use PVSCSI disks and multiple virtual machines using vSphere FT are running on the same host.
This issue is resolved in this release.
- PR 1988248: An ESXi host might fail with a purple diagnostic screen due to filesystem journal allocation issue
An ESXi host might fail with a purple diagnostic screen if the filesystem journal allocation fails while opening a volume due to space limitation or I/O timeout, and the volume opens as read only.
This issue is resolved in this release.
- PR 2024647: An ESXi host might fail with a purple diagnostic screen and an error message when you try to retrieve information about multicast group that contains more than 128 IP addresses
When you add more than 128 multicast IP addresses to an ESXi host and try to retrieve information about the multicast group, the ESXi host might fail with a purple diagnostic screen and an error message such as:
PANIC bora/vmkernel/main/dlmalloc.c:4924 - Usage error in dlmalloc
.This issue is resolved in this release.
- PR 2039373: Hostd might fail due to a high rate of Map Disk Region tasks generated by a backup solution
Backup solutions might produce a high rate of Map Disk Region tasks, causing hostd to run out of memory and fail.
This issue is resolved in this release.
- PR 1972487: An ESXi host might fail with a purple diagnostic screen due to a rare race condition
An ESXi host might fail with a purple diagnostic screen due to a rare race condition between file allocation paths and volume extension operations.
This issue is resolved in this release.
- PR 1913198: NFS 4.1 mounts might fail when using AUTH_SYS, KRB5 or KRB5i authentication
If you use AUTH_SYS, KRB5 or KRB5i authentication to mount NFS 4.1 shares, the mount might fail. This happens when the Generic Security Services type is not the same value for all the NFS requests generated during the mount session. Some hosts, which are less tolerant of such inconsistencies, fail the mount requests.
This issue is resolved in this release.
- PR 1882261: ESXi hosts might report wrong esxcli storage core device list data
For some devices, such as CD-ROM, where the maximum queue depth is less than the default value of number of outstanding I/Os with competing worlds, the command
esxcli storage core device list
might show the maximum number of outstanding I/O requests exceeding the maximum queue depth. With this fix, the maximum value ofNo of outstanding IOs with competing worlds
is restricted to the maximum queue depth of the storage device on which you make the change.This issue is resolved in this release.
- PR 1824576: ESXi hosts might lose connectivity during snapshot scalability testing
The Small-Footprint CIM Broker (SFCB) might fail during snapshot operations on large scale with a core dump file similar to
sfcb-vmware_bas-zdump.XXX
in/var/core/
and cause ESXi hosts to become unresponsive.This issue is resolved in this release. - PR 2032048: An ESXi host might fail with an error due to invalid keyboard input
An ESXi host might fail with an error
kbdmode_set:519: invalid keyboard mode 4: Not supported
due to an invalid keyboard input during the shutdown process.This issue is resolved in this release.
- PR 1937646: The sfcb-hhrc process might stop working, causing the hardware monitoring service on the ESXi host to become unavailable
The
sfcb-hhrc
process might stop working, causing the hardware monitoring service sfcbd on the ESXi host to become unavailable.This issue is resolved in this release.
- PR 2016411: When you add an ESXi host to an Active Directory domain, hostd might become unresponsive and the ESXi host disconnects from vCenter Server
When you add an ESXi host to an Active Directory domain, hostd might become unresponsive due to Likewise services running out of memory and the ESXi host disconnects from the vCenter Server system.
This issue is resolved in this release.
- PR 1954364: Virtual machines with EFI firmware might fail to boot from Windows Deployment Services on Windows Server 2008
PXE boot of virtual machines with 64-bit Extensible Firmware Interface (EFI) might fail when booting from Windows Deployment Services on Windows Server 2008, because PXE boot discovery requests might contain an incorrect PXE architecture ID. This fix sets a default architecture ID in line with the processor architecture identifier for x64 UEFI that the Internet Engineering Task Force (IETF) stipulates. This release also adds a new advanced configuration option,
efi.pxe.architectureID = <integer>
, to facilitate backwards compatibility. For instance, if your environment is configured to use architecture ID 9, you can addefi.pxe.architectureID = 9
to the advanced configuration options of the virtual machine.This issue is resolved in this release.
- PR 1962739: Shared raw device mapping devices might go offline after a vMotion operation with MSCS clustered virtual machines
If a backend failover of a SAN storage controller takes place after a vMotion operation of MSCS clustered virtual machines, shared raw device mapping devices might go offline.
The issue is resolved in this release.
- PR 2046087: An ESXi host might fail with a purple diagnostic screen and NULL pointer value
An ESXi host might fail with a purple diagnostic screen in the process of collecting information for statistical purposes due to stale TCP connections.
This issue is resolved in this release.
- PR 1986020: ESXi VMFS5 and VMFS6 file systems might fail to expand
ESXi VMFS5 and VMFS6 file systems with certain volume sizes might fail to expand due to an issue with the logical volume manager (LVM). A VMFS6 volume fails to expand with a size greater than 16 TB.
This issue is resolved in this release.
- PR 1977715: ESXi hosts intermittently stop sending logs to remote syslog servers
The Syslog Service of ESXi hosts might stop transferring logs to remote log servers when a remote server is down and does not restart transfers after the server is up.
This issue is resolved in this release.
- PR 1998437: SATP ALUA might not handle properly SCSI sense code 0x2 0x4 0xa
SATP ALUA might not handle properly SCSI sense code 0x2 0x4 0xa, which translates to the alert
LOGICAL UNIT NOT ACCESSIBLE, ASYMMETRIC ACCESS STATE TRANSITION
. This causes a premature all paths down state that might lead to VMFS heartbeat timeouts and ultimately to performance drops. With this fix, you can use the advanced configuration option/Nmp/Nmp Satp Alua Cmd RetryTime
and increase the timeout value to 50 seconds.This issue is resolved in this release.
- PR 2058697: EtherSwitch service modules might run out of memory when configuring a L3 mirror session
EtherSwitch service modules might run out of memory when configuring a L3 mirror session with Encapsulated Remote Switched Port Analyzer (ERSPAN) type II or type III.
This issue is resolved in this release.
- PR 1996632: ESXi hosts might intermittently lose connectivity if configured to use an Active Directory domain with a large number of accounts
If an ESXi host is configured to use an Active Directory domain with more than 10,000 accounts, the host might intermittently lose connectivity.
This issue is resolved in this release.
- PR 1995640: An ESXi host might fail with a purple diagnostic screen with PCPU locked error
An ESXi host might fail with a purple diagnostic screen with an error for locked PCPUs, caused by a spin count fault.
You might see a log similar to this:
2017-08-05T21:51:50.370Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T21:56:55.398Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:02:00.418Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:07:05.438Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:12:10.453Z cpu2:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:17:15.472Z cpu2:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T16:32:08.539Z cpu67:67618 opID=da84f05c)FSS: 6229: Conflict between buffered and unbuffered open (file 'DC_UAT_NETSERVER-000001-sesparse.vmdk'):flags 0x4002, requested flags 0x8
2017-08-05T16:32:45.177Z cpu61:67618 opID=389fe17f)FSS: 6229: Conflict between buffered and unbuffered open (file 'DC_UAT_Way4SRV-000001-sesparse.vmdk'):flags 0x4002, requested flags 0x8
2017-08-05T16:33:16.529Z cpu66:68466 opID=db8b1ddc)FSS: 6229: Conflict between buffered and unbuffered open (file 'DC_UAT_FIleServer-000001-sesparse.vmdk'):flags 0x4002, requested flags 0x8
This issue is resolved in this release.
- PR 2053426: Virtual machines with more than 1TB of memory and PCI passthrough devices might fail to power on
The VMkernel module that manages PCI passthrough devices has a dedicated memory heap that does not allow support to virtual machines configured with more than 1 TB of memory and one or more PCI passthrough devices. Such virtual machines might fail to power on with a warning in the ESXi vmkernel log (
/var/log/vmkernel.log
) similar toWARNING: Heap: 3729: Heap mainHeap (2618200/31462232): Maximum allowed growth (28844032) too small for size (32673792)
. The issue also occurs when multiple virtual machines, each with less than 1 TB of memory and one or more PCI passthrough devices, are powered on concurrently and the sum of the memory configured across all virtual machines exceeds 1 TB.This issue is resolved in this release.
- PR 1987996: Warning appears in vSAN health service Hardware compatibility check: Timed out in querying HCI info
When vSAN health service runs a Hardware compatibility check, the following warning message might appear:
Timed out in querying HCI info.
This warning is caused by slow retrieval of hardware information. The check times out after 20 seconds, which is less than the time required to retrieve information from the disks.This issue is resolved in this release.
- PR 2048702: ImageConfigManager causes increased log activity in the syslog.log file
The
syslog.log
file is repeatedly populated with messages related to ImageConfigManager.This issue is resolved in this release.
- PR 2051775: The Common Information Model interface of an ESXi host might stop working and fail hardware health monitoring
You might not be able to monitor the hardware health of an ESXi host through the CIM interface, because some processes in SFCB might stop working. This fix provides code enhancements to the sfcb-hhrc and the sfcb-vmware_base processes. The
sfcb-hhrc process
manages third-party Host Hardware RAID Controller (HHRC) CIM providers and provides hardware health monitoring, and thesfcb-vmware_base
process.This issue is resolved in this release.
- PR 2055746: HotAdd backup of a Windows virtual machine disk might fail if both the target virtual machine and the proxy virtual machine are enabled for Content Based Read Cache (CBRC)
HotAdd backup of a Windows virtual machine disk might fail if both the target virtual machine and the proxy virtual machine are CBRC-enabled, because the snapshot disk of the target virtual machine cannot be added to the backup proxy virtual machine.
This issue is resolved in this release.
- PR 2057918: VMkernel logs in VMware vSAN environment might be flooded with the message Unable to register file system
In vSAN environments, VMkernel logs that record activities related to virtual machines and ESXi, might be flooded with the message
Unable to register file system
.This issue is resolved in this release.
- PR 2052300: Attempts to create a host profile might fail with an error due to a null value in iSCSI configuration
Attempts to create a host profile might fail with an error
NoneType
object has no attributelower
if you pass a null value parameter in the iSCSI configuration.This issue resolved in this release.
- PR 2057189: Packet loss across L2 VPN networks
You might see packet loss of up to 15% and poor application performance across a L2 VPN network if a sink port is configured on the vSphere Distributed Switch but no overlay network is available.
This issue is resolved in this release.
- PR 1911721: The esxtop utility reports incorrect statistics for DAVG/cmd and KAVG/cmd on VAAI-supported LUNs
The
esxtop
utility reports incorrect statistics for the average device latency per command (DAVG/cmd) and average ESXi VMkernel latency per command (KAVG/cmd) on VAAI-supported LUNs due to an incorrect calculation.This issue is resolved in this release.
- PR 1678559: You might notice performance degradation after resizing a VMDK file on vSAN
Distributed Object Manager processing operations might take too much time when an object has concatenated components. This might cause delays in other operations, which leads to high I/O latency.
This issue is resolved in this release.
- PR 2064616: Application running parallel File I/O operations might cause the ESXi host to stop responding
Application running parallel File I/O operations on a VMFS datastore might cause the ESXi host to stop responding.
This issue is resolved in this release.
- PR 2034930: When you customize a Host Profile some parameter values might be lost
When you customize a Host Profile to attach it to a host, some parameter values, including host name and MAC address, might be lost.
This issue is resolved in this release.
- PR 2034625: An ESXi host might fail with a purple diagnostic screen due to faulty packets from third-party networking modules
If VMware vSphere Network I/O Control is not enabled, packets from third-party networking modules might cause an ESXi host to fail with a purple diagnostic screen.
This issue is resolved in this release.
- PR 1994373: You might fail to gather data to diagnose networking issues
With ESXi 6.5 Update 2, you can use PacketCapture, a lightweight tcpdump utility implemented in the Reverse HTTP Proxy service, to capture and store only the minimum amount of data to diagnose a networking issue, saving CPU and storage. For more information, see KB 52843.
This issue is resolved in this release.
- PR 1950800: VMkernel adapter fails to communicate through an Emulex physical NIC after an ESXi host boots
After an ESXi host boots, the VMkernel adapter that connects to virtual switches with an Emulex physical NIC might be unable to communicate though the NIC and the VMkernel adapter might not get a DHCP IP address.
This issue is resolved in this release.
- The ESXCLI unmap command might reclaim only small file block resources in VMFS-6 datastores and not process large file block resources
In VMFS-6 datastores, if you use the
esxcli storage vmfs unmap
command for manual reclamation of free space, it might only reclaim space from small file blocks and not process large file block resources.This issue is resolved in this release.
- PR 2060589: vSphere Web Client might display incorrect storage allocation after an increase of the disk size of a virtual machine
If you use vSphere Web Client to increase the disk size of a virtual machine, the vSphere Web Client might not fully reflect the reconfiguration and display the storage allocation of the virtual machine as unchanged.
This issue is resolved in this release.
- PR 2012602: An ESXi host might not have access to the Web Services for Management (WSMan) protocol due to failed CIM ticket authentication
An ESXi host might not have access to the WSMan protocol due to failing CIM ticket authentication.
This issue is resolved in this release.
- PR 2065000: Hostd might fail during backup
The ESXi daemon hostd might fail during backup due to a race condition.
This issue is resolved in this release.
- PR 2015082: Deleting files from a content library might cause the vCenter Server Agent to fail
Deleting files from a content library might cause a failure of the vCenter Server Agent (vpxa) service which results in failure of the delete operation.
This issue is resolved in this release.
- PR 1997995: Dell system management software might not display the BIOS attributes of ESXi hosts on Dell OEM servers
Dell system management software might not display the BIOS attributes of ESXi hosts on Dell OEM servers, because the ESXi kernel module fails to load automatically during ESXi boot. This is due to different vendor names in the SMBIOS System Information of Dell and Dell OEM servers.
This issue is resolved in this release.
- PR 1491595: Hostd might fail due to a memory leak
Under certain circumstances, the memory of hostd might exceed the hard limit and the agent might fail.
This issue is resolved in this release.
- PR 2034772: An ESXi host might fail with a purple screen due to a rare race condition in iSCSI sessions
An ESXi host might fail with a purple diagnostic screen due to a rare race condition in iSCSI sessions while processing the scheduled queue and running queue during connection shutdown. The connection lock might be freed to terminate a task while the shutdown is still in process. The next node pointer would be incorrect and cause the failure.
This issue resolved in this release.
- PR 2056969: Log congestion leads to unresponsive hosts and inaccessible virtual machines
If a device in a disk group encounters a permanent disk loss, the physical log (PLOG) in the cache device might increase to its maximum threshold. This problem is due to faulty cleanup of data in the log, which are bound to the lost disk. The log entries build up over time, leading to log congestion, which can cause virtual machines to become inaccessible.
This issue is resolved in this release.
- PR 2054940: vSAN health service unable to read controller information
If you are using HP controllers, vSAN health service might be unable to read the controller firmware information. The health check might issue a warning, with the firmware version listed as N/A.
This issue is resolved in this release.
- PR 2034800: Ping failure in vSAN network health check when VMware vRealize Operations manages the vSAN cluster
When vSAN cluster is managed by vRealize Operations, you might observe intermittent ping health check failures.
This issue is resolved in this release.
- PR 2034452: The vSAN health service might be slow displaying data for a large scale cluster after integration with vRealize Operations
After vRealize Operations integration with vSAN Management Pack enabled, the vSAN health service might become less responsive. This problem affects large scale clusters, such as those with 24 hosts, each with 10 disks. vSAN health service uses an inefficient log filter for parsing some specific log patterns, and can consume a large amount of CPU resources while communicating with vRealize Operations.
This issue is resolved in this release.
- PR 2023740: Host might fail with a purple diagnostic screen error if its own local configuration is added through esxcli vsan cluster unicastagent add
When adding unicast configuration of cluster hosts, if the host own local configuration is added, the host experiences a purple diagnostic screen error.
This issue is resolved in this release.
- PR 2033471: Very low IOPS on vSAN iSCSI target when outstanding I/O is 1
For each iSCSI target, the initiator establishes one or more connections depending on whether multipath is configured. For each connection, if there is only one outstanding I/O all the time, and the network is very fast (ping time less than 1 ms), the IOPS on an iSCSI LUN might be much lower that on a native vSAN LUN.
This issue is resolved in this release.
- PR 2028736: A host in a vSAN cluster might fail with a page fault
When the instantiation of a vSAN object fails due to memory constraints, the host might experience a purple diagnostic screen error.
This issue is resolved in this release.
- PR 2023858: vSAN health service alarm could delay 5 to 10 minutes to inform hosts with connection issue
After a host has lost connection with vCenter Server, with no manual retest, there could be a 5 to 10 minute delay for vSAN health service to report the loss of connection.
This issue is resolved in this release.
- PR 2021819: False alarm for cluster configuration consistency health check after enabling vSAN encryption
A false alarm is reported for cluster configuration consistency health check when vSAN encryption is enabled with two or more KMS server configured. The KMS certifications fetched from host and vCenter Server do not match due to improper ordering.
This issue is resolved in this release.
- PR 2059300: Host failure when upgrading to vCenter Server 6.5 Update 2
During upgrade to vCenter Server 6.5 Update 2, when vSAN applies unicast configuration of other hosts in the cluster, a local configuration might be added on the host, resulting in a host failure with purple diagnostic screen.
This issue is resolved in this release.
- PR 2020615: vSphere vMotion operations for virtual machines with more than 64GB of vRAM might fail on vSAN
When you perform a vSphere vMotion operation with a virtual machine with more than 64GB of memory (vRAM) on vSAN, the task might fail due to error in swap file initialization on the destination.
This issue is resolved in this release.
- PR 2008663: After reboot, host is delayed from joining cluster running vSAN 6.6 and later
When there are many vSAN components on disks, during reboot it might take a long time to publish vSAN storage. The host is delayed from joining the vSAN cluster, so this host might be partitioned for a long time.
This issue is resolved in this release.
- PR 2023766: Large amounts of resync traffic for an object being repaired or placed in maintenance mode
When an object spans across all available fault domains, vSAN is unable to replace a component that is impacted by a maintenance mode operation or a repair operation. If this happens, vSAN rebuilds the entire object, which can result in large amounts of resync traffic.
This issue is resolved in this release.
- PR 2039896: Some operations fail and the host must be rebooted
When vSAN Observer is running, memory leaks can occur in the host init group. Other operations that run under the init group might fail until you reboot the host.
This issue is resolved in this release.
- PR 1999439: Slow performance and client latency due to excess DOM I/O activity
Under certain conditions, when I/Os from the Distributed Object Manager (DOM) fail and are retried, the cluster might experience high latencies and slow performance. You might see the following trace logs:
[33675821] [cpu66] [6bea734c OWNER writeWithBlkAttr5 VMDISK] DOMTraceOperationNeedsRetry:3630: {'op': 0x43a67f18c6c0, 'obj': 0x439e90cb0c00, 'objUuid': '79ec2259-fd98-ef85-0e1a-1402ec44eac0', 'status': 'VMK_STORAGE_RETRY_OPERATION'}
This issue is resolved in this release.
- PR 2061175: I/O performance of the input–output memory management unit (IOMMU) driver of AMD might be slow on ESX PCIe devices
I/O performance of ESX PCIe devices might drop when the IOMMU driver of AMD is enabled.
This issue is resolved in this release.
- PR 1982291: Virtual machines with brackets in their names might not deploy successfully
Virtual machines with brackets ([]) in their name might not deploy successfully. For more information, see KB 53648.
This issue is resolved in this release.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
Related CVE numbers | N/A |
This patch updates the qedentv VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 1998108 |
Related CVE numbers | N/A |
This patch updates the vmkusb VIB to resolve the following issue:
- PR 1998108: An ESXi host might fail with a purple diagnostic screen and a Page Fault error message when you attach a USB device
An ESXi host might fail with a purple diagnostic screen and a
Page Fault
error message when you attach a USB device. This happens because some USB devices report invalid number of interface descriptors to the host.This issue is resolved in this release.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 1948936 |
Related CVE numbers | N/A |
This patch updates ipmi-ipmi-devintf and ipmi-ipmi-msghandler VIBs to resolve the following issue:
- PR 1948936: Third-party CIM providers or agents might cause memory corruptions that lead to destabilization of the kernel or the drivers of an ESXi host
Third-party CIM providers or agents might cause memory corruptions that lead to destabilization of the kernel or the drivers of an ESXi host and the host might fail with a purple diagnostic screen. Memory corruptions are visible at later point, and not at the exact time of corruption.
This issue is resolved in this release.
Patch Category | Enhancement |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2004950 |
Related CVE numbers | N/A |
This patch updates the vmw-ahci VIB to resolve the following issue:
- PR 2004950: Update to the VMware Advanced Host Controller Interface driver
The VMware Advanced Host Controller Interface (AHCI) driver is updated to version 1.1.1-1 to fix issues with memory allocation, the Marvell 9230 AHCI controller, and Intel Cannon Lake PCH-H SATA AHCI controller.
Patch Category | Enhancement |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2004948 |
Related CVE numbers | N/A |
This patch updates the nhpsa VIB to resolve the following issue:
- PR 2004948: Update to the nhpsa driver
The nhpsa driver is updated to 2.0.22-1vmw.
Patch Category | Enhancement |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2004949 |
Related CVE numbers | N/A |
This patch updates the nvme VIB to resolve the following issue:
- PR 2004949: Updates to the NVMe driver
ESXi 6.5 Update 2 adds management of multiple namespaces compatible with the NVMe 1.2 specification and enhanced diagnostic log.
Patch Category | Enhancement |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2004957 |
Related CVE numbers | N/A |
The VIB in this bulletin is new and you might see the following expected behavior:
- The command esxcli software vib update -d <zip bundle> might skip the new VIB. You must use the command esxcli software profile update --depot=depot_location --profile=profile_name instead.
- After running the command esxcli software vib update -n <vib name>, in Installation Result you might see the message Host not changed. You must use the command esxcli software vib install -n instead.
This patch updates the smartpqi VIB to resolve the following issue:
- PR 2004957: smartpqi driver for HPE Smart PQI controllers
ESXi 6.5 Update 2 enables smartpqi driver support for the HPE ProLiant Gen10 Smart Array Controller.
Patch Category | Enhancement |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2004956 |
Related CVE numbers | N/A |
The VIB in this bulletin is new and you might see the following expected behavior:
- The command esxcli software vib update -d <zip bundle> might skip the new VIB. You must use the command esxcli software profile update --depot=depot_location --profile=profile_name instead.
- After running the command esxcli software vib update -n <vib name>, in Installation Result you might see the message Host not changed. You must use the command esxcli software vib install -n instead.
This patch updates the lsi-msgpt35 VIB to resolve the following issue:
- PR 2004956: I/O devices enablement for Broadcom SAS 3.5 IT/IR controllers
ESXi 6.5 Update 2 enables a new native driver to support the Broadcom SAS 3.5 IT/IR controllers with devices including combination of NVMe, SAS, and SATA drives. The HBA device IDs are: SAS3516(00AA), SAS3516_1(00AB), SAS3416(00AC), SAS3508(00AD), SAS3508_1(00AE), SAS3408(00AF), SAS3716(00D0), SAS3616(00D1), SAS3708(00D2).
Patch Category | Enhancement |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2004947 |
Related CVE numbers | N/A |
This patch updates the lsi-msgpt3 VIB to resolve the following issue:
- PR 2004947: Update to the lsi_msgpt3 driver
The lsi_msgpt3 driver is updated to version 16.00.01.00-1vmw.
Patch Category | Enhancement |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2004946 |
Related CVE numbers | N/A |
This patch updates the lsi-mr3 VIB to resolve the following issue:
- PR 2004946: Update to the lsi_mr3 driver to version MR 7.2
The
lsi_mr3
driver is updated to enable support for Broadcom SAS 3.5 RAID controllers. It also contains significant enhancement on task management and critical bug fixes.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
Related CVE numbers | N/A |
The VIB in this bulletin is new and you might see the following expected behavior:
- The command esxcli software vib update -d <zip bundle> might skip the new VIB. You must use the command esxcli software profile update --depot=depot_location --profile=profile_name instead.
- After running the command esxcli software vib update -n <vib name>, in Installation Result you might see the message Host not changed. You must use the command esxcli software vib install -n instead.
This patch updates the bnxtnet VIB.
Patch Category | Enhancement |
Patch Severity | Important |
Host Reboot Required | No |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2004949 |
Related CVE numbers | N/A |
This patch updates the vmware-esx-esxcli-nvme-plugin VIB to resolve the following issue:
- PR 2004949: Updates to the NVMe driver
ESXi 6.5 Update 2 adds management of multiple namespaces compatible with the NVMe 1.2 specification and enhanced diagnostic log.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
Related CVE numbers | N/A |
This patch updates the i40en VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
Related CVE numbers | N/A |
This patch updates the misc-drivers VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2015606 |
Related CVE numbers | N/A |
This patch updates the lsi-msgpt2 VIB to resolve the following issue:
- PR 2015606: VMware ESXi native drivers might fail to support more than eight message signaled interrupts
ESXi native drivers might fail to support more than eight message signaled interrupts even on devices with MSI-X that allows the device to allocate up to 2048 interrupts. This issue is observed on newer AMD-based platforms and is due to default limits of the native drivers, lsi_msgpt2 and lsi_msgpt3.
This issue is resolved in this release.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
Related CVE numbers | N/A |
This patch updates the ne1000 VIB.
Patch Category | Enhancement |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
Related CVE numbers | N/A |
This patch updates the ixgben VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | No |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2036299 |
Related CVE numbers | N/A |
This patch updates the lsu-hp-hpsa-plugin VIB.
- PR 2036299: The ESXi native driver for HPE Smart Array controllers extends the list of expander devices to enable compatibility with HPE Gen9 configurations
The disk serviceability plug-in of the ESXi native driver for HPE Smart Array controllers, nhpsa, now works with an extended list of expander devices to enable compatibility with HPE Gen9 configurations.
Patch Category | Bugfix |
Patch Severity | Moderate |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2053073 |
Related CVE numbers | N/A |
This patch updates the lpfc VIB to resolve the following issue:
- PR 2053073: Support for LightPulse and OneConnect adapters is split to separate default drivers
In ESXi 6.5 Update 2, LightPulse and OneConnect adapters are supported by separate default drivers. The brcmfcoe driver supports OneConnect adapters and the lpfc driver supports only LightPulse adapters. Previously, the lpfc driver supported both OneConnect and LightPulse adapters.
Patch Category | Bugfix |
Patch Severity | Moderate |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2053073 |
Related CVE numbers | N/A |
The VIB in this bulletin is new and you might see the following expected behavior:
- The command esxcli software vib update -d <zip bundle> might skip the new VIB. You must use the command esxcli software profile update --depot=depot_location --profile=profile_name instead.
- After running the command esxcli software vib update -n <vib name>, in Installation Result you might see the message Host not changed. You must use the command esxcli software vib install -n instead.
This patch updates the brcmfcoe VIB to resolve the following issue:
- PR 2053073: Support for LightPulse and OneConnect adapters is split to separate default drivers
In ESXi 6.5 Update 2, LightPulse and OneConnect adapters are supported by separate default drivers. The brcmfcoe driver supports OneConnect adapters and the lpfc driver supports only LightPulse adapters. Previously, the lpfc driver supported both OneConnect and LightPulse adapters.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2053073 |
Related CVE numbers | N/A |
This patch updates the usbcore-usb VIB to resolve the following issue:
- PR 2053703: An ESXi host might fail with purple diagnostic screen when completing installation from a USB media drive on HP servers
When you press Enter to reboot an ESXi host and complete a new installation from a USB media drive on some HP servers, you might see an error on a purple diagnostic screen due to a rare race condition.
This issue is resolved in this release.
Patch Category | Bugfix |
Patch Severity | Critical |
Host Reboot Required | No |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2051280 |
Related CVE numbers | N/A |
This patch updates the esx-xserver VIB to resolve the following issue:
- PR 2051280: Some Single Root I/O Virtualization (SR-IOV) graphics configurations that are running ESXi 6.5 with enabled GPU Statistics might cause virtual machines to become unresponsive
Some SR-IOV graphics configurations, that are running ESXi 6.5 with enabled GPU Statistics , might cause virtual machines on the ESXi host to become unresponsive after extended host uptime due to a memory leak. In some cases, the ESXi host might become unresponsive as well.
This issue is resolved in this release.
Patch Category | Security |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 1827139, 2011288, 1964481 |
Related CVE numbers | N/A |
This patch updates esx-base, esx-tboot,vsan and vsanhealth VIBs to resolve the following issues:
- PR 1827139: ESXi hosts might randomly fail due to errors in sfcbd or other daemon processes
This fix improves internal error handling to prevent ESXi host failures, caused by sfcbd or other daemon processes.
This issue is resolved in this release.
- PR 2011288: Update to glibc package to address security issues
The glibc package has been updated to address CVE-2017-15804, CVE-2017-15670, CVE-2016-4429, CVE-2015-8779, CVE-2015-8778, CVE-2014-9761 and CVE-2015-8776.
- PR 1964481: Update of the SQLite database
The SQLite database is updated to version 3.20.1 to resolve a security issue with identifier CVE-2017-10989.
Patch Category | Security |
Patch Severity | Important |
Host Reboot Required | No |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
Related CVE numbers | N/A |
This patch updates the tools-light VIB.
Patch Category | Security |
Patch Severity | Important |
Host Reboot Required | No |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
Related CVE numbers | N/A |
This patch updates the esx-ui VIB.
Profile Name | ESXi-6.5.0-20180502001-standard |
Build | For build information, see Patches Contained in this Release. |
Vendor | VMware, Inc. |
Release Date | May 3, 2018 |
Acceptance Level | PartnerSupported |
Affected Hardware | N/A |
Affected Software | N/A |
Affected VIBs |
|
PRs Fixed | 1882261, 1907025, 1911721, 1913198, 1913483, 1919300, 1923173, 1937646, 1947334, 1950336, 1950800, 1954364, 1958600, 1962224, 1962739, 1971168, 1972487, 1975787, 1977715, 1982291, 1983123, 1986020, 1987996, 1988248, 1995640, 1996632, 1997995, 1998437, 1999439, 2004751, 2006929, 2007700, 2008663, 2011392, 2012602, 2015132, 2016411, 2016413, 2018046, 2020615, 2020785, 2021819, 2023740, 2023766, 2023858, 2024647, 2028736, 2029476, 2032048, 2033471, 2034452, 2034625, 2034772, 2034800, 2034930, 2039373, 2039896, 2046087, 2048702, 2051775, 2052300, 2053426, 2054940, 2055746, 2057189, 2057918, 2058697, 2059300, 2060589, 2061175, 2064616, 2065000, 1998108, 1948936, 2004950, 2004948, 2004949, 2004957, 2004956, 2004947, 1491595, 2004946, 2004949, 2015606, 2036299, 2053073, 2053073, 2053703, 2051280, 1994373 |
Related CVE numbers | N/A |
- This patch updates the following issues:
- PR 1913483: Third-party storage array user interfaces might not properly display relationships between virtual machine disks and virtual machines
Third-party storage array user interfaces might not properly display relationships between virtual machine disks and virtual machines when one disk is attached to many virtual machines, due to insufficient metadata.
This issue is resolved in this release.
- PR 1975787: An ESXi support bundle generated by using the VMware Host Client might not contain all log files
Sometimes, when you generate a support bundle for an ESXi host by using the Host Client, the support bundle might not contain all log files.
This issue is resolved in this release.
- PR 1947334: Attempts to modify the vSAN iSCSI parameter VSAN-iSCSI.ctlWorkerThreads fail
You cannot modify the vSAN iSCSI parameter
VSAN-iSCSI.ctlWorkerThreads
, because the parameter is designed as an internal parameter and is not a subject to change. This fix hides internal parameters of vSAN iSCSI to avoid confusion.This issue is resolved in this release.
- PR 1919300: Host Profile compliance check might show Unknown status if a profile is imported from a different vCenter Server system
If you export a Host Profile from one vCenter Server system and import it to another vCenter Server system, compliance checks might show a status Unknown.
This issue is resolved in this release.
- PR 2007700: The advanced setting VMFS.UnresolvedVolumeLiveCheck might be inactive in the vSphere Web Client
In the vSphere Web Client, the option to enable or disable checks during unresolved volume queries with the VMFS.UnresolvedVolumeLiveCheck option under Advance Setting > VMFS, might be grayed out.
This issue is resolved in this release. You can also workaround it by using the vSphere Client or set the option by directly connecting to the ESXi host.
- PR 1983123: ESXi hosts with scratch partitions located on a vSAN datastore might intermittently become unresponsive
If the scratch partition of an ESXi host is located on a vSAN datastore, the host might become unresponsive, because if the datastore is temporary inaccessible, this might block userworld daemons, such as hostd, if they try to access this datastore at the same time. With this fix, scratch partitions to a vSAN datastore are no longer configurable through the API or the user interface.
This issue is resolved in this release.
- PR 1950336: Hostd might stop responding in attempt to reconfigure a virtual machine
Hostd might stop responding in attempt to reconfigure virtual machines. Configuration entries with unset value in the ConfigSpec object, which encapsulates configuration settings for creation and configuration of virtual machines, cause the issue.
This issue is resolved in this release.
- PR 2006929: VMware vSphere vMotion might fail due to timeout and inaccessible virtual machines
vSphere vMotion might fail due to timeout, because virtual machines might need more buffer to store the shared memory page details used by 3D or graphic devices. The issue is observed mainly on virtual machines running graphics intensive loads or working in virtual desktop infrastructure environments. With this fix, the buffer size is increased by four times for 3D virtual machines and if the increased capacity is still not enough, the virtual machine is powered off.
This issue is resolved in this release.
- PR 1958600: The netdump IP address might be lost after upgrading from ESXi 5.5 Update 3b to ESXi 6.0 Update 3 and/or ESXi 6.5 Update 2
The netdump IP address might be lost after your upgrade from ESXi 5.5 Update 3b to ESXi 6.0 Update 3 and/or ESXi 6.5 Update 2.
This issue is resolved in this release.
- PR 1971168: ESXi system identification information in the vSphere Web Client might not be available
System identification information consists of asset tags, service tags, and OEM strings. In earlier releases, this information comes from the Common Information Model (CIM) service. Starting with vSphere 6.5, the CIM service is turned off by default, unless any third-party CIM providers are installed, which prevents ESXi system identification information to display in the vSphere Web Client.
This issue is resolved in this release.
- PR 2020785: Virtual machines created by the VMware Integrated OpenStack might not work properly
When you create virtual machine disks with the VMware Integrated OpenStack block storage service (Cinder), the disks get random UUIDs. In around 0.5% of the cases, the virtual machines do not recognize these UUIDs, which results in unexpected behavior of the virtual machines.
This issue is resolved in this release.
- PR 2004751: A virtual machine might fail to boot after an upgrade to macOS 10.13
When you upgrade a virtual machine to macOS 10.13, the virtual machine might fail to locate a bootable operating system. If the macOS 10.13 installer determines that the system volume is located on a Solid State Disk (SSD), during the OS upgrade, the installer converts the Hierarchical File System Extended (HFS+/HFSX) to the new Apple File System (APFS) format. As a result, the virtual machine might fail to boot.
This issue is resolved in this release.
- PR 2015132: You might not be able to apply a Host Profile
You might not be able to apply a host profile, if the host profile contains a DNS configuration for a VMkernel network adapter that is associated to a vSphere Distributed Switch.
This issue is resolved in this release.
- PR 2011392: An ESXi host might fail with a purple diagnostic screen due to a deadlock in vSphere Distributed Switch health checks
If you enable VLAN and maximum transmission unit (MTU) checks in vSphere Distributed Switches, your ESXi hosts might fail with a purple diagnostic screen due to a possible deadlock.
This issue is resolved in this release.
- PR 2004948: Update to the nhpsa driver
The nhpsa driver is updated to 2.0.22-1vmw.
- PR 1907025: An ESXi host might fail with an error Unexpected message type
An ESXi host might fail with a purple diagnostic screen due to invalid network input in the error message handling of vSphere Fault Tolerance, which results into an
Unexpected message type
error.This issue is resolved in this release.
- PR 1962224: Virtual machines might fail with an error Unexpected FT failover due to PANIC
When you run vSphere Fault Tolerance in vSphere 6.5 and vSphere 6.0, virtual machines might fail with an error such as Unexpected FT failover due to PANIC error: VERIFY bora/vmcore/vmx/main/monitorAction.c:598. The error occurs when virtual machines use PVSCSI disks and multiple virtual machines using vSphere FT are running on the same host.
This issue is resolved in this release.
- PR 1988248: An ESXi host might fail with a purple diagnostic screen due to filesystem journal allocation issue
An ESXi host might fail with a purple diagnostic screen if the filesystem journal allocation fails while opening a volume due to space limitation or I/O timeout, and the volume opens as read only.
This issue is resolved in this release.
- PR 2024647: An ESXi host might fail with a purple diagnostic screen and an error message when you try to retrieve information about multicast group that contains more than 128 IP addresses
When you add more than 128 multicast IP addresses to an ESXi host and try to retrieve information about the multicast group, the ESXi host might fail with a purple diagnostic screen and an error message such as:
PANIC bora/vmkernel/main/dlmalloc.c:4924 - Usage error in dlmalloc
.This issue is resolved in this release.
- PR 2039373: Hostd might fail due to a high rate of Map Disk Region tasks generated by a backup solution
Backup solutions might produce a high rate of Map Disk Region tasks, causing hostd to run out of memory and fail.
This issue is resolved in this release.
- PR 1972487: An ESXi host might fail with a purple diagnostic screen due to a rare race condition
An ESXi host might fail with a purple diagnostic screen due to a rare race condition between file allocation paths and volume extension operations.
This issue is resolved in this release.
- PR 1913198: NFS 4.1 mounts might fail when using AUTH_SYS, KRB5 or KRB5i authentication
If you use AUTH_SYS, KRB5 or KRB5i authentication to mount NFS 4.1 shares, the mount might fail. This happens when the Generic Security Services type is not the same value for all the NFS requests generated during the mount session. Some hosts, which are less tolerant of such inconsistencies, fail the mount requests.
This issue is resolved in this release.
- PR 1882261: ESXi hosts might report wrong esxcli storage core device list data
For some devices, such as CD-ROM, where the maximum queue depth is less than the default value of number of outstanding I/Os with competing worlds, the command
esxcli storage core device list
might show the maximum number of outstanding I/O requests exceeding the maximum queue depth. With this fix, the maximum value ofNo of outstanding IOs with competing worlds
is restricted to the maximum queue depth of the storage device on which you make the change.This issue is resolved in this release.
- PR 1824576: ESXi hosts might lose connectivity during snapshot scalability testing
The Small-Footprint CIM Broker (SFCB) might fail during snapshot operations on large scale with a core dump file similar to
sfcb-vmware_bas-zdump.XXX
in/var/core/
and cause ESXi hosts to become unresponsive.This issue is resolved in this release. - PR 2032048: An ESXi host might fail with an error due to invalid keyboard input
An ESXi host might fail with an error
kbdmode_set:519: invalid keyboard mode 4: Not supported
due to an invalid keyboard input during the shutdown process.This issue is resolved in this release.
- PR 1937646: The sfcb-hhrc process might stop working, causing the hardware monitoring service on the ESXi host to become unavailable
The
sfcb-hhrc
process might stop working, causing the hardware monitoring service sfcbd on the ESXi host to become unavailable.This issue is resolved in this release.
- PR 2016411: When you add an ESXi host to an Active Directory domain, hostd might become unresponsive and the ESXi host disconnects from vCenter Server
When you add an ESXi host to an Active Directory domain, hostd might become unresponsive due to Likewise services running out of memory and the ESXi host disconnects from the vCenter Server system.
This issue is resolved in this release.
- PR 1954364: Virtual machines with EFI firmware might fail to boot from Windows Deployment Services on Windows Server 2008
PXE boot of virtual machines with 64-bit Extensible Firmware Interface (EFI) might fail when booting from Windows Deployment Services on Windows Server 2008, because PXE boot discovery requests might contain an incorrect PXE architecture ID. This fix sets a default architecture ID in line with the processor architecture identifier for x64 UEFI that the Internet Engineering Task Force (IETF) stipulates. This release also adds a new advanced configuration option,
efi.pxe.architectureID = <integer>
, to facilitate backwards compatibility. For instance, if your environment is configured to use architecture ID 9, you can addefi.pxe.architectureID = 9
to the advanced configuration options of the virtual machine.This issue is resolved in this release.
- PR 1962739: Shared raw device mapping devices might go offline after a vMotion operation with MSCS clustered virtual machines
If a backend failover of a SAN storage controller takes place after a vMotion operation of MSCS clustered virtual machines, shared raw device mapping devices might go offline.
The issue is resolved in this release.
- PR 2046087: An ESXi host might fail with a purple diagnostic screen and NULL pointer value
An ESXi host might fail with a purple diagnostic screen in the process of collecting information for statistical purposes due to stale TCP connections.
This issue is resolved in this release.
- PR 1986020: ESXi VMFS5 and VMFS6 file systems might fail to expand
ESXi VMFS5 and VMFS6 file systems with certain volume sizes might fail to expand due to an issue with the logical volume manager (LVM). A VMFS6 volume fails to expand with a size greater than 16 TB.
This issue is resolved in this release.
- PR 1977715: ESXi hosts intermittently stop sending logs to remote syslog servers
The Syslog Service of ESXi hosts might stop transferring logs to remote log servers when a remote server is down and does not restart transfers after the server is up.
This issue is resolved in this release.
- PR 1998437: SATP ALUA might not handle properly SCSI sense code 0x2 0x4 0xa
SATP ALUA might not handle properly SCSI sense code 0x2 0x4 0xa, which translates to the alert
LOGICAL UNIT NOT ACCESSIBLE, ASYMMETRIC ACCESS STATE TRANSITION
. This causes a premature all paths down state that might lead to VMFS heartbeat timeouts and ultimately to performance drops. With this fix, you can use the advanced configuration option/Nmp/Nmp Satp Alua Cmd RetryTime
and increase the timeout value to 50 seconds.This issue is resolved in this release.
- PR 2058697: EtherSwitch service modules might run out of memory when configuring a L3 mirror session
EtherSwitch service modules might run out of memory when configuring a L3 mirror session with Encapsulated Remote Switched Port Analyzer (ERSPAN) type II or type III.
This issue is resolved in this release.
- PR 1996632: ESXi hosts might intermittently lose connectivity if configured to use an Active Directory domain with a large number of accounts
If an ESXi host is configured to use an Active Directory domain with more than 10,000 accounts, the host might intermittently lose connectivity.
This issue is resolved in this release.
- PR 1995640: An ESXi host might fail with a purple diagnostic screen with PCPU locked error
An ESXi host might fail with a purple diagnostic screen with an error for locked PCPUs, caused by a spin count fault.
You might see a log similar to this:
2017-08-05T21:51:50.370Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T21:56:55.398Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:02:00.418Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:07:05.438Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:12:10.453Z cpu2:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:17:15.472Z cpu2:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T16:32:08.539Z cpu67:67618 opID=da84f05c)FSS: 6229: Conflict between buffered and unbuffered open (file 'DC_UAT_NETSERVER-000001-sesparse.vmdk'):flags 0x4002, requested flags 0x8
2017-08-05T16:32:45.177Z cpu61:67618 opID=389fe17f)FSS: 6229: Conflict between buffered and unbuffered open (file 'DC_UAT_Way4SRV-000001-sesparse.vmdk'):flags 0x4002, requested flags 0x8
2017-08-05T16:33:16.529Z cpu66:68466 opID=db8b1ddc)FSS: 6229: Conflict between buffered and unbuffered open (file 'DC_UAT_FIleServer-000001-sesparse.vmdk'):flags 0x4002, requested flags 0x8
This issue is resolved in this release.
- PR 2053426: Virtual machines with more than 1TB of memory and PCI passthrough devices might fail to power on
The VMkernel module that manages PCI passthrough devices has a dedicated memory heap that does not allow support to virtual machines configured with more than 1 TB of memory and one or more PCI passthrough devices. Such virtual machines might fail to power on with a warning in the ESXi vmkernel log (
/var/log/vmkernel.log
) similar toWARNING: Heap: 3729: Heap mainHeap (2618200/31462232): Maximum allowed growth (28844032) too small for size (32673792)
. The issue also occurs when multiple virtual machines, each with less than 1 TB of memory and one or more PCI passthrough devices, are powered on concurrently and the sum of the memory configured across all virtual machines exceeds 1 TB.This issue is resolved in this release.
- PR 1987996: Warning appears in vSAN health service Hardware compatibility check: Timed out in querying HCI info
When vSAN health service runs a Hardware compatibility check, the following warning message might appear:
Timed out in querying HCI info.
This warning is caused by slow retrieval of hardware information. The check times out after 20 seconds, which is less than the time required to retrieve information from the disks.This issue is resolved in this release.
- PR 2048702: ImageConfigManager causes increased log activity in the syslog.log file
The
syslog.log
file is repeatedly populated with messages related to ImageConfigManager.This issue is resolved in this release.
- PR 2051775: The Common Information Model interface of an ESXi host might stop working and fail hardware health monitoring
You might not be able to monitor the hardware health of an ESXi host through the CIM interface, because some processes in SFCB might stop working. This fix provides code enhancements to the sfcb-hhrc and the sfcb-vmware_base processes. The
sfcb-hhrc process
manages third-party Host Hardware RAID Controller (HHRC) CIM providers and provides hardware health monitoring, and thesfcb-vmware_base
process.This issue is resolved in this release.
- PR 2055746: HotAdd backup of a Windows virtual machine disk might fail if both the target virtual machine and the proxy virtual machine are enabled for Content Based Read Cache (CBRC)
HotAdd backup of a Windows virtual machine disk might fail if both the target virtual machine and the proxy virtual machine are CBRC-enabled, because the snapshot disk of the target virtual machine cannot be added to the backup proxy virtual machine.
This issue is resolved in this release.
- PR 2057918: VMkernel logs in VMware vSAN environment might be flooded with the message Unable to register file system
In vSAN environments, VMkernel logs that record activities related to virtual machines and ESXi, might be flooded with the message
Unable to register file system
.This issue is resolved in this release.
- PR 2052300: Attempts to create a host profile might fail with an error due to a null value in iSCSI configuration
Attempts to create a host profile might fail with an error
NoneType
object has no attributelower
if you pass a null value parameter in the iSCSI configuration.This issue resolved in this release.
- PR 2057189: Packet loss across L2 VPN networks
You might see packet loss of up to 15% and poor application performance across a L2 VPN network if a sink port is configured on the vSphere Distributed Switch but no overlay network is available.
This issue is resolved in this release.
- PR 1911721: The esxtop utility reports incorrect statistics for DAVG/cmd and KAVG/cmd on VAAI-supported LUNs
The
esxtop
utility reports incorrect statistics for the average device latency per command (DAVG/cmd) and average ESXi VMkernel latency per command (KAVG/cmd) on VAAI-supported LUNs due to an incorrect calculation.This issue is resolved in this release.
- PR 1678559: You might notice performance degradation after resizing a VMDK file on vSAN
Distributed Object Manager processing operations might take too much time when an object has concatenated components. This might cause delays in other operations, which leads to high I/O latency.
This issue is resolved in this release.
- PR 2064616: Application running parallel File I/O operations might cause the ESXi host to stop responding
Application running parallel File I/O operations on a VMFS datastore might cause the ESXi host to stop responding.
This issue is resolved in this release.
- PR 2034930: When you customize a Host Profile some parameter values might be lost
When you customize a Host Profile to attach it to a host, some parameter values, including host name and MAC address, might be lost.
This issue is resolved in this release.
- PR 2034625: An ESXi host might fail with a purple diagnostic screen due to faulty packets from third-party networking modules
If VMware vSphere Network I/O Control is not enabled, packets from third-party networking modules might cause an ESXi host to fail with a purple diagnostic screen.
This issue is resolved in this release.
- PR 1994373: You might fail to gather data to diagnose networking issues
With ESXi 6.5 Update 2, you can use PacketCapture, a lightweight tcpdump utility implemented in the Reverse HTTP Proxy service, to capture and store only the minimum amount of data to diagnose a networking issue, saving CPU and storage. For more information, see KB 52843.
This issue is resolved in this release.
- PR 1950800: VMkernel adapter fails to communicate through an Emulex physical NIC after an ESXi host boots
After an ESXi host boots, the VMkernel adapter that connects to virtual switches with an Emulex physical NIC might be unable to communicate though the NIC and the VMkernel adapter might not get a DHCP IP address.
This issue is resolved in this release.
- PR 2060589: vSphere Web Client might display incorrect storage allocation after an increase of the disk size of a virtual machine
If you use vSphere Web Client to increase the disk size of a virtual machine, the vSphere Web Client might not fully reflect the reconfiguration and display the storage allocation of the virtual machine as unchanged.
This issue is resolved in this release.
- PR 2012602: An ESXi host might not have access to the Web Services for Management (WSMan) protocol due to failed CIM ticket authentication
An ESXi host might not have access to the WSMan protocol due to failing CIM ticket authentication.
This issue is resolved in this release.
- PR 2065000: Hostd might fail during backup
The ESXi daemon hostd might fail during backup due to a race condition.
This issue is resolved in this release.
- PR 2015082: Deleting files from a content library might cause the vCenter Server Agent to fail
Deleting files from a content library might cause a failure of the vCenter Server Agent (vpxa) service which results in failure of the delete operation.
This issue is resolved in this release.
- PR 1997995: Dell system management software might not display the BIOS attributes of ESXi hosts on Dell OEM servers
Dell system management software might not display the BIOS attributes of ESXi hosts on Dell OEM servers, because the ESXi kernel module fails to load automatically during ESXi boot. This is due to different vendor names in the SMBIOS System Information of Dell and Dell OEM servers.
This issue is resolved in this release.
- PR 1491595: Hostd might fail due to a memory leak
Under certain circumstances, the memory of hostd might exceed the hard limit and the agent might fail.
This issue is resolved in this release.
- PR 2034772: An ESXi host might fail with a purple screen due to a rare race condition in iSCSI sessions
An ESXi host might fail with a purple diagnostic screen due to a rare race condition in iSCSI sessions while processing the scheduled queue and running queue during connection shutdown. The connection lock might be freed to terminate a task while the shutdown is still in process. The next node pointer would be incorrect and cause the failure.
This issue resolved in this release.
- PR 205696: Log congestion leads to unresponsive hosts and inaccessible virtual machines
If a device in a disk group encounters a permanent disk loss, the physical log (PLOG) in the cache device might increase to its maximum threshold. This problem is due to faulty cleanup of data in the log, which are bound to the lost disk. The log entries build up over time, leading to log congestion, which can cause virtual machines to become inaccessible.
This issue is resolved in this release.
- PR 2054940: vSAN health service unable to read controller information
If you are using HP controllers, vSAN health service might be unable to read the controller firmware information. The health check might issue a warning, with the firmware version listed as N/A.
This issue is resolved in this release.
- PR 2034800: Ping failure in vSAN network health check when VMware vRealize Operations manages the vSAN cluster
When vSAN cluster is managed by vRealize Operations, you might observe intermittent ping health check failures.
This issue is resolved in this release.
- PR 2034452: The vSAN health service might be slow displaying data for a large scale cluster after integration with vRealize Operations
After vRealize Operations integration with vSAN Management Pack enabled, the vSAN health service might become less responsive. This problem affects large scale clusters, such as those with 24 hosts, each with 10 disks. vSAN health service uses an inefficient log filter for parsing some specific log patterns, and can consume a large amount of CPU resources while communicating with vRealize Operations.
This issue is resolved in this release.
- PR 2023740: Host might fail with a purple diagnostic screen error if its own local configuration is added through esxcli vsan cluster unicastagent add
When adding unicast configuration of cluster hosts, if the host own local configuration is added, the host experiences a purple diagnostic screen error.
This issue is resolved in this release.
- PR 2033471: Very low IOPS on vSAN iSCSI target when outstanding I/O is 1
For each iSCSI target, the initiator establishes one or more connections depending on whether multipath is configured. For each connection, if there is only one outstanding I/O all the time, and the network is very fast (ping time less than 1 ms), the IOPS on an iSCSI LUN might be much lower that on a native vSAN LUN.
This issue is resolved in this release.
- PR 2028736: A host in a vSAN cluster might fail with a page fault
When the instantiation of a vSAN object fails due to memory constraints, the host might experience a purple diagnostic screen error.
This issue is resolved in this release.
- PR 2023858: vSAN health service alarm could delay 5 to 10 minutes to inform hosts with connection issue
After a host has lost connection with vCenter Server, with no manual retest, there could be a 5 to 10 minute delay for vSAN health service to report the loss of connection.
This issue is resolved in this release.
- PR 2021819: False alarm for cluster configuration consistency health check after enabling vSAN encryption
A false alarm is reported for cluster configuration consistency health check when vSAN encryption is enabled with two or more KMS server configured. The KMS certifications fetched from host and vCenter Server do not match due to improper ordering.
This issue is resolved in this release.
- PR 2059300: Host failure when upgrading to vCenter Server 6.5 Update 2
During upgrade to vCenter Server 6.5 Update 2, when vSAN applies unicast configuration of other hosts in the cluster, a local configuration might be added on the host, resulting in a host failure with purple diagnostic screen.
This issue is resolved in this release.
- PR 2020615: vSphere vMotion operations for virtual machines with more than 64GB of vRAM might fail on vSAN
When you perform a vSphere vMotion operation with a virtual machine with more than 64GB of memory (vRAM) on vSAN, the task might fail due to error in swap file initialization on the destination.
This issue is resolved in this release.
- PR 2008663: After reboot, host is delayed from joining cluster running vSAN 6.6 and later
When there are many vSAN components on disks, during reboot it might take a long time to publish vSAN storage. The host is delayed from joining the vSAN cluster, so this host might be partitioned for a long time.
This issue is resolved in this release.
- PR 2023766: Large amounts of resync traffic for an object being repaired or placed in maintenance mode
When an object spans across all available fault domains, vSAN is unable to replace a component that is impacted by a maintenance mode operation or a repair operation. If this happens, vSAN rebuilds the entire object, which can result in large amounts of resync traffic.
This issue is resolved in this release.
- PR 2039896: Some operations fail and the host must be rebooted
When vSAN Observer is running, memory leaks can occur in the host init group. Other operations that run under the init group might fail until you reboot the host.
This issue is resolved in this release.
- PR 1999439: Slow performance and client latency due to excess DOM I/O activity
Under certain conditions, when I/Os from the Distributed Object Manager (DOM) fail and are retried, the cluster might experience high latencies and slow performance. You might see the following trace logs:
[33675821] [cpu66] [6bea734c OWNER writeWithBlkAttr5 VMDISK] DOMTraceOperationNeedsRetry:3630: {'op': 0x43a67f18c6c0, 'obj': 0x439e90cb0c00, 'objUuid': '79ec2259-fd98-ef85-0e1a-1402ec44eac0', 'status': 'VMK_STORAGE_RETRY_OPERATION'}
This issue is resolved in this release.
- PR 2061175: I/O performance of the input–output memory management unit (IOMMU) driver of AMD might be slow on ESX PCIe devices
I/O performance of ESX PCIe devices might drop when the IOMMU driver of AMD is enabled.
This issue is resolved in this release.
- PR 1982291: Virtual machines with brackets in their names might not deploy successfully
Virtual machines with brackets ([]) in their name might not deploy successfully. For more information, see KB 53648.
This issue is resolved in this release.
-
PR 1950800: VMkernel adapter fails to communicate through an Emulex physical NIC after an ESXi host boots
After an ESXi host boots, the VMkernel adapter that connects to virtual switches with an Emulex physical NIC might be unable to communicate though the NIC and the VMkernel adapter might not get a DHCP IP address.
Workaround: You must unplug the LAN cable and plug it again.
- PR 1998108: An ESXi host might fail with a purple diagnostic screen and a Page Fault error message when you attach a USB device
An ESXi host might fail with a purple diagnostic screen and a
Page Fault
error message when you attach a USB device. This happens because some USB devices report invalid number of interface descriptors to the host.This issue is resolved in this release.
- PR 1948936: Third-party CIM providers or agents might cause memory corruptions that lead to destabilization of the kernel or the drivers of an ESXi host
Third-party CIM providers or agents might cause memory corruptions that lead to destabilization of the kernel or the drivers of an ESXi host and the host might fail with a purple diagnostic screen. Memory corruptions are visible at later point, and not at the exact time of corruption.
This issue is resolved in this release.
- PR 2004950: Update to the VMware Advanced Host Controller Interface driver
The VMware Advanced Host Controller Interface (AHCI) driver is updated to version 1.1.1-1 to fix issues with memory allocation, the Marvell 9230 AHCI controller, and Intel Cannon Lake PCH-H SATA AHCI controller.
- PR 2004948: Update to the nhpsa driver
The nhpsa driver is updated to 2.0.22-1vmw.
- PR 2004949: Updates to the NVMe driver
ESXi 6.5 Update 2 adds management of multiple namespaces compatible with the NVMe 1.2 specification and enhanced diagnostic log.
- PR 2004957: smartpqi driver for HPE Smart PQI controllers
ESXi 6.5 Update 2 enables smartpqi driver support for the HPE ProLiant Gen10 Smart Array Controller.
- PR 2004956: I/O devices enablement for Broadcom SAS 3.5 IT/IR controllers
ESXi 6.5 Update 2 enables a new native driver to support the Broadcom SAS 3.5 IT/IR controllers with devices including combination of NVMe, SAS, and SATA drives. The HBA device IDs are: SAS3516(00AA), SAS3516_1(00AB), SAS3416(00AC), SAS3508(00AD), SAS3508_1(00AE), SAS3408(00AF), SAS3716(00D0), SAS3616(00D1), SAS3708(00D2).
- PR 2004947: Update to the lsi_msgpt3 driver
The lsi_msgpt3 driver is updated to version 16.00.01.00-1vmw.
- PR 2004946: Update to the lsi_mr3 driver to version MR 7.2
The
lsi_mr3
driver is updated to enable support for Broadcom SAS 3.5 RAID controllers. It also contains significant enhancement on task management and critical bug fixes. - PR 2004949: Updates to the NVMe driver
ESXi 6.5 Update 2 adds management of multiple namespaces compatible with the NVMe 1.2 specification and enhanced diagnostic log.
- PR 2015606: VMware ESXi native drivers might fail to support more than eight message signaled interrupts
ESXi native drivers might fail to support more than eight message signaled interrupts even on devices with MSI-X that allows the device to allocate up to 2048 interrupts. This issue is observed on newer AMD-based platforms and is due to default limits of the native drivers, lsi_msgpt2 and lsi_msgpt3.
This issue is resolved in this release.
- PR 2036299: The ESXi native driver for HPE Smart Array controllers extends the list of expander devices to enable compatibility with HPE Gen9 configurations
The disk serviceability plug-in of the ESXi native driver for HPE Smart Array controllers, nhpsa, now works with an extended list of expander devices to enable compatibility with HPE Gen9 configurations.
- PR 2053073: Support for LightPulse and OneConnect adapters is split to separate default drivers
In ESXi 6.5 Update 2, LightPulse and OneConnect adapters are supported by separate default drivers. The brcmfcoe driver supports OneConnect adapters and the lpfc driver supports only LightPulse adapters. Previously, the lpfc driver supported both OneConnect and LightPulse adapters.
- PR 2053703: An ESXi host might fail with purple diagnostic screen when completing installation from a USB media drive on HP servers
When you press Enter to reboot an ESXi host and complete a new installation from a USB media drive on some HP servers, you might see an error on a purple diagnostic screen due to a rare race condition.
This issue is resolved in this release.
- PR 2051280: Some Single Root I/O Virtualization (SR-IOV) graphics configurations that are running ESXi 6.5 with enabled GPU Statistics might cause virtual machines to become unresponsive
Some SR-IOV graphics configurations, that are running ESXi 6.5 with enabled GPU Statistics , might cause virtual machines on the ESXi host to become unresponsive after extended host uptime due to a memory leak. In some cases, the ESXi host might become unresponsive as well.
This issue is resolved in this release.
- The ESXCLI unmap command might reclaim only small file block resources in VMFS-6 datastores and not process large file block resources
In VMFS-6 datastores, if you use the
esxcli storage vmfs unmap
command for manual reclamation of free space, it might only reclaim space from small file blocks and not process large file block resources.This issue is resolved in this release.
- PR 1950800: VMkernel adapter fails to communicate through an Emulex physical NIC after an ESXi host boots
After an ESXi host boots, the VMkernel adapter that connects to virtual switches with an Emulex physical NIC might be unable to communicate though the NIC and the VMkernel adapter might not get a DHCP IP address.
Workaround: You must unplug the LAN cable and plug it again.
- PR 1913483: Third-party storage array user interfaces might not properly display relationships between virtual machine disks and virtual machines
Profile Name | ESXi-6.5.0-20180502001-standard |
Build | For build information, see Patches Contained in this Release. |
Vendor | VMware, Inc. |
Release Date | May 3, 2018 |
Acceptance Level | PartnerSupported |
Affected Hardware | N/A |
Affected Software | N/A |
Affected VIBs |
|
PRs Fixed | 1882261, 1907025, 1911721, 1913198, 1913483, 1919300, 1923173, 1937646, 1947334, 1950336, 1950800, 1954364, 1958600, 1962224, 1962739, 1971168, 1972487, 1975787, 1977715, 1982291, 1983123, 1986020, 1987996, 1988248, 1995640, 1996632, 1997995, 1998437, 1999439, 2004751, 2006929, 2007700, 2008663, 2011392, 2012602, 2015132, 2016411, 2016413, 2018046, 2020615, 2020785, 2021819, 2023740, 2023766, 2023858, 2024647, 2028736, 2029476, 2032048, 2033471, 2034452, 2034625, 2034772, 2034800, 2034930, 2039373, 2039896, 2046087, 2048702, 2051775, 2052300, 2053426, 2054940, 2055746, 2057189, 2057918, 2058697, 2059300, 2060589, 2061175, 2064616, 2065000, 1998108, 1948936, 2004950, 2004948, 2004949, 2004957, 2004956, 2004947, 1491595, 2004946, 2004949, 2015606, 2036299, 2053073, 2053073, 2053703, 2051280, 1994373 |
Related CVE numbers | N/A |
- This patch updates the following issues:
- PR 1913483: Third-party storage array user interfaces might not properly display relationships between virtual machine disks and virtual machines
Third-party storage array user interfaces might not properly display relationships between virtual machine disks and virtual machines when one disk is attached to many virtual machines, due to insufficient metadata.
This issue is resolved in this release.
- PR 1975787: An ESXi support bundle generated by using the VMware Host Client might not contain all log files
Sometimes, when you generate a support bundle for an ESXi host by using the Host Client, the support bundle might not contain all log files.
This issue is resolved in this release.
- PR 1947334: Attempts to modify the vSAN iSCSI parameter VSAN-iSCSI.ctlWorkerThreads fail
You cannot modify the vSAN iSCSI parameter
VSAN-iSCSI.ctlWorkerThreads
, because the parameter is designed as an internal parameter and is not a subject to change. This fix hides internal parameters of vSAN iSCSI to avoid confusion.This issue is resolved in this release.
- PR 1919300: Host Profile compliance check might show Unknown status if a profile is imported from a different vCenter Server system
If you export a Host Profile from one vCenter Server system and import it to another vCenter Server system, compliance checks might show a status Unknown.
This issue is resolved in this release.
- PR 2007700: The advanced setting VMFS.UnresolvedVolumeLiveCheck might be inactive in the vSphere Web Client
In the vSphere Web Client, the option to enable or disable checks during unresolved volume queries with the VMFS.UnresolvedVolumeLiveCheck option under Advance Setting > VMFS, might be grayed out.
This issue is resolved in this release. You can also workaround it by using the vSphere Client or set the option by directly connecting to the ESXi host.
- PR 1983123: ESXi hosts with scratch partitions located on a vSAN datastore might intermittently become unresponsive
If the scratch partition of an ESXi host is located on a vSAN datastore, the host might become unresponsive, because if the datastore is temporary inaccessible, this might block userworld daemons, such as hostd, if they try to access this datastore at the same time. With this fix, scratch partitions to a vSAN datastore are no longer configurable through the API or the user interface.
This issue is resolved in this release.
- PR 1950336: Hostd might stop responding in attempt to reconfigure a virtual machine
Hostd might stop responding in attempt to reconfigure virtual machines. Configuration entries with unset value in the ConfigSpec object, which encapsulates configuration settings for creation and configuration of virtual machines, cause the issue.
This issue is resolved in this release.
- PR 2006929: VMware vSphere vMotion might fail due to timeout and inaccessible virtual machines
vSphere vMotion might fail due to timeout, because virtual machines might need more buffer to store the shared memory page details used by 3D or graphic devices. The issue is observed mainly on virtual machines running graphics intensive loads or working in virtual desktop infrastructure environments. With this fix, the buffer size is increased by four times for 3D virtual machines and if the increased capacity is still not enough, the virtual machine is powered off.
This issue is resolved in this release.
- PR 1958600: The netdump IP address might be lost after upgrading from ESXi 5.5 Update 3b to ESXi 6.0 Update 3 and/or ESXi 6.5 Update 2
The netdump IP address might be lost after your upgrade from ESXi 5.5 Update 3b to ESXi 6.0 Update 3 and/or ESXi 6.5 Update 2.
This issue is resolved in this release.
- PR 1971168: ESXi system identification information in the vSphere Web Client might not be available
System identification information consists of asset tags, service tags, and OEM strings. In earlier releases, this information comes from the Common Information Model (CIM) service. Starting with vSphere 6.5, the CIM service is turned off by default, unless any third-party CIM providers are installed, which prevents ESXi system identification information to display in the vSphere Web Client.
This issue is resolved in this release.
- PR 2020785: Virtual machines created by the VMware Integrated OpenStack might not work properly
When you create virtual machine disks with the VMware Integrated OpenStack block storage service (Cinder), the disks get random UUIDs. In around 0.5% of the cases, the virtual machines do not recognize these UUIDs, which results in unexpected behavior of the virtual machines.
This issue is resolved in this release.
- PR 2004751: A virtual machine might fail to boot after an upgrade to macOS 10.13
When you upgrade a virtual machine to macOS 10.13, the virtual machine might fail to locate a bootable operating system. If the macOS 10.13 installer determines that the system volume is located on a Solid State Disk (SSD), during the OS upgrade, the installer converts the Hierarchical File System Extended (HFS+/HFSX) to the new Apple File System (APFS) format. As a result, the virtual machine might fail to boot.
This issue is resolved in this release.
- PR 2015132: You might not be able to apply a Host Profile
You might not be able to apply a host profile, if the host profile contains a DNS configuration for a VMkernel network adapter that is associated to a vSphere Distributed Switch.
This issue is resolved in this release.
- PR 2011392: An ESXi host might fail with a purple diagnostic screen due to a deadlock in vSphere Distributed Switch health checks
If you enable VLAN and maximum transmission unit (MTU) checks in vSphere Distributed Switches, your ESXi hosts might fail with a purple diagnostic screen due to a possible deadlock.
This issue is resolved in this release.
- PR 2004948: Update to the nhpsa driver
The nhpsa driver is updated to 2.0.22-1vmw.
- PR 1907025: An ESXi host might fail with an error Unexpected message type
An ESXi host might fail with a purple diagnostic screen due to invalid network input in the error message handling of vSphere Fault Tolerance, which results into an
Unexpected message type
error.This issue is resolved in this release.
- PR 1962224: Virtual machines might fail with an error Unexpected FT failover due to PANIC
When you run vSphere Fault Tolerance in vSphere 6.5 and vSphere 6.0, virtual machines might fail with an error such as Unexpected FT failover due to PANIC error: VERIFY bora/vmcore/vmx/main/monitorAction.c:598. The error occurs when virtual machines use PVSCSI disks and multiple virtual machines using vSphere FT are running on the same host.
This issue is resolved in this release.
- PR 1988248: An ESXi host might fail with a purple diagnostic screen due to filesystem journal allocation issue
An ESXi host might fail with a purple diagnostic screen if the filesystem journal allocation fails while opening a volume due to space limitation or I/O timeout, and the volume opens as read only.
This issue is resolved in this release.
- PR 2024647: An ESXi host might fail with a purple diagnostic screen and an error message when you try to retrieve information about multicast group that contains more than 128 IP addresses
When you add more than 128 multicast IP addresses to an ESXi host and try to retrieve information about the multicast group, the ESXi host might fail with a purple diagnostic screen and an error message such as:
PANIC bora/vmkernel/main/dlmalloc.c:4924 - Usage error in dlmalloc
.This issue is resolved in this release.
- PR 2039373: Hostd might fail due to a high rate of Map Disk Region tasks generated by a backup solution
Backup solutions might produce a high rate of Map Disk Region tasks, causing hostd to run out of memory and fail.
This issue is resolved in this release.
- PR 1972487: An ESXi host might fail with a purple diagnostic screen due to a rare race condition
An ESXi host might fail with a purple diagnostic screen due to a rare race condition between file allocation paths and volume extension operations.
This issue is resolved in this release.
- PR 1913198: NFS 4.1 mounts might fail when using AUTH_SYS, KRB5 or KRB5i authentication
If you use AUTH_SYS, KRB5 or KRB5i authentication to mount NFS 4.1 shares, the mount might fail. This happens when the Generic Security Services type is not the same value for all the NFS requests generated during the mount session. Some hosts, which are less tolerant of such inconsistencies, fail the mount requests.
This issue is resolved in this release.
- PR 1882261: ESXi hosts might report wrong esxcli storage core device list data
For some devices, such as CD-ROM, where the maximum queue depth is less than the default value of number of outstanding I/Os with competing worlds, the command
esxcli storage core device list
might show the maximum number of outstanding I/O requests exceeding the maximum queue depth. With this fix, the maximum value ofNo of outstanding IOs with competing worlds
is restricted to the maximum queue depth of the storage device on which you make the change.This issue is resolved in this release.
- PR 1824576: ESXi hosts might lose connectivity during snapshot scalability testing
The Small-Footprint CIM Broker (SFCB) might fail during snapshot operations on large scale with a core dump file similar to
sfcb-vmware_bas-zdump.XXX
in/var/core/
and cause ESXi hosts to become unresponsive.This issue is resolved in this release. - PR 2032048: An ESXi host might fail with an error due to invalid keyboard input
An ESXi host might fail with an error
kbdmode_set:519: invalid keyboard mode 4: Not supported
due to an invalid keyboard input during the shutdown process.This issue is resolved in this release.
- PR 1937646: The sfcb-hhrc process might stop working, causing the hardware monitoring service on the ESXi host to become unavailable
The
sfcb-hhrc
process might stop working, causing the hardware monitoring service sfcbd on the ESXi host to become unavailable.This issue is resolved in this release.
- PR 2016411: When you add an ESXi host to an Active Directory domain, hostd might become unresponsive and the ESXi host disconnects from vCenter Server
When you add an ESXi host to an Active Directory domain, hostd might become unresponsive due to Likewise services running out of memory and the ESXi host disconnects from the vCenter Server system.
This issue is resolved in this release.
- PR 1954364: Virtual machines with EFI firmware might fail to boot from Windows Deployment Services on Windows Server 2008
PXE boot of virtual machines with 64-bit Extensible Firmware Interface (EFI) might fail when booting from Windows Deployment Services on Windows Server 2008, because PXE boot discovery requests might contain an incorrect PXE architecture ID. This fix sets a default architecture ID in line with the processor architecture identifier for x64 UEFI that the Internet Engineering Task Force (IETF) stipulates. This release also adds a new advanced configuration option,
efi.pxe.architectureID = <integer>
, to facilitate backwards compatibility. For instance, if your environment is configured to use architecture ID 9, you can addefi.pxe.architectureID = 9
to the advanced configuration options of the virtual machine.This issue is resolved in this release.
- PR 1962739: Shared raw device mapping devices might go offline after a vMotion operation with MSCS clustered virtual machines
If a backend failover of a SAN storage controller takes place after a vMotion operation of MSCS clustered virtual machines, shared raw device mapping devices might go offline.
The issue is resolved in this release.
- PR 2046087: An ESXi host might fail with a purple diagnostic screen and NULL pointer value
An ESXi host might fail with a purple diagnostic screen in the process of collecting information for statistical purposes due to stale TCP connections.
This issue is resolved in this release.
- PR 1986020: ESXi VMFS5 and VMFS6 file systems might fail to expand
ESXi VMFS5 and VMFS6 file systems with certain volume sizes might fail to expand due to an issue with the logical volume manager (LVM). A VMFS6 volume fails to expand with a size greater than 16 TB.
This issue is resolved in this release.
- PR 1977715: ESXi hosts intermittently stop sending logs to remote syslog servers
The Syslog Service of ESXi hosts might stop transferring logs to remote log servers when a remote server is down and does not restart transfers after the server is up.
This issue is resolved in this release.
- PR 1998437: SATP ALUA might not handle properly SCSI sense code 0x2 0x4 0xa
SATP ALUA might not handle properly SCSI sense code 0x2 0x4 0xa, which translates to the alert
LOGICAL UNIT NOT ACCESSIBLE, ASYMMETRIC ACCESS STATE TRANSITION
. This causes a premature all paths down state that might lead to VMFS heartbeat timeouts and ultimately to performance drops. With this fix, you can use the advanced configuration option/Nmp/Nmp Satp Alua Cmd RetryTime
and increase the timeout value to 50 seconds.This issue is resolved in this release.
- PR 2058697: EtherSwitch service modules might run out of memory when configuring a L3 mirror session
EtherSwitch service modules might run out of memory when configuring a L3 mirror session with Encapsulated Remote Switched Port Analyzer (ERSPAN) type II or type III.
This issue is resolved in this release.
- PR 1996632: ESXi hosts might intermittently lose connectivity if configured to use an Active Directory domain with a large number of accounts
If an ESXi host is configured to use an Active Directory domain with more than 10,000 accounts, the host might intermittently lose connectivity.
This issue is resolved in this release.
- PR 1995640: An ESXi host might fail with a purple diagnostic screen with PCPU locked error
An ESXi host might fail with a purple diagnostic screen with an error for locked PCPUs, caused by a spin count fault.
You might see a log similar to this:
2017-08-05T21:51:50.370Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T21:56:55.398Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:02:00.418Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:07:05.438Z cpu4:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:12:10.453Z cpu2:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T22:17:15.472Z cpu2:134108)WARNING: Hbr: 873: Failed to receive from 172.16.30.42 (groupID=GID-d6bb98fb-ace7-4793-9ee5-30545518a5a3): Broken pipe
2017-08-05T16:32:08.539Z cpu67:67618 opID=da84f05c)FSS: 6229: Conflict between buffered and unbuffered open (file 'DC_UAT_NETSERVER-000001-sesparse.vmdk'):flags 0x4002, requested flags 0x8
2017-08-05T16:32:45.177Z cpu61:67618 opID=389fe17f)FSS: 6229: Conflict between buffered and unbuffered open (file 'DC_UAT_Way4SRV-000001-sesparse.vmdk'):flags 0x4002, requested flags 0x8
2017-08-05T16:33:16.529Z cpu66:68466 opID=db8b1ddc)FSS: 6229: Conflict between buffered and unbuffered open (file 'DC_UAT_FIleServer-000001-sesparse.vmdk'):flags 0x4002, requested flags 0x8
This issue is resolved in this release.
- PR 2053426: Virtual machines with more than 1TB of memory and PCI passthrough devices might fail to power on
The VMkernel module that manages PCI passthrough devices has a dedicated memory heap that does not allow support to virtual machines configured with more than 1 TB of memory and one or more PCI passthrough devices. Such virtual machines might fail to power on with a warning in the ESXi vmkernel log (
/var/log/vmkernel.log
) similar toWARNING: Heap: 3729: Heap mainHeap (2618200/31462232): Maximum allowed growth (28844032) too small for size (32673792)
. The issue also occurs when multiple virtual machines, each with less than 1 TB of memory and one or more PCI passthrough devices, are powered on concurrently and the sum of the memory configured across all virtual machines exceeds 1 TB.This issue is resolved in this release.
- PR 1987996: Warning appears in vSAN health service Hardware compatibility check: Timed out in querying HCI info
When vSAN health service runs a Hardware compatibility check, the following warning message might appear:
Timed out in querying HCI info.
This warning is caused by slow retrieval of hardware information. The check times out after 20 seconds, which is less than the time required to retrieve information from the disks.This issue is resolved in this release.
- PR 2048702: ImageConfigManager causes increased log activity in the syslog.log file
The
syslog.log
file is repeatedly populated with messages related to ImageConfigManager.This issue is resolved in this release.
- PR 2051775: The Common Information Model interface of an ESXi host might stop working and fail hardware health monitoring
You might not be able to monitor the hardware health of an ESXi host through the CIM interface, because some processes in SFCB might stop working. This fix provides code enhancements to the sfcb-hhrc and the sfcb-vmware_base processes. The
sfcb-hhrc process
manages third-party Host Hardware RAID Controller (HHRC) CIM providers and provides hardware health monitoring, and thesfcb-vmware_base
process.This issue is resolved in this release.
- PR 2055746: HotAdd backup of a Windows virtual machine disk might fail if both the target virtual machine and the proxy virtual machine are enabled for Content Based Read Cache (CBRC)
HotAdd backup of a Windows virtual machine disk might fail if both the target virtual machine and the proxy virtual machine are CBRC-enabled, because the snapshot disk of the target virtual machine cannot be added to the backup proxy virtual machine.
This issue is resolved in this release.
- PR 2057918: VMkernel logs in VMware vSAN environment might be flooded with the message Unable to register file system
In vSAN environments, VMkernel logs that record activities related to virtual machines and ESXi, might be flooded with the message
Unable to register file system
.This issue is resolved in this release.
- PR 2052300: Attempts to create a host profile might fail with an error due to a null value in iSCSI configuration
Attempts to create a host profile might fail with an error
NoneType
object has no attributelower
if you pass a null value parameter in the iSCSI configuration.This issue resolved in this release.
- PR 2057189: Packet loss across L2 VPN networks
You might see packet loss of up to 15% and poor application performance across a L2 VPN network if a sink port is configured on the vSphere Distributed Switch but no overlay network is available.
This issue is resolved in this release.
- PR 1911721: The esxtop utility reports incorrect statistics for DAVG/cmd and KAVG/cmd on VAAI-supported LUNs
The
esxtop
utility reports incorrect statistics for the average device latency per command (DAVG/cmd) and average ESXi VMkernel latency per command (KAVG/cmd) on VAAI-supported LUNs due to an incorrect calculation.This issue is resolved in this release.
- PR 1678559: You might notice performance degradation after resizing a VMDK file on vSAN
Distributed Object Manager processing operations might take too much time when an object has concatenated components. This might cause delays in other operations, which leads to high I/O latency.
This issue is resolved in this release.
- PR 2064616: Application running parallel File I/O operations might cause the ESXi host to stop responding
Application running parallel File I/O operations on a VMFS datastore might cause the ESXi host to stop responding.
This issue is resolved in this release.
- PR 2034930: When you customize a Host Profile some parameter values might be lost
When you customize a Host Profile to attach it to a host, some parameter values, including host name and MAC address, might be lost.
This issue is resolved in this release.
- PR 2034625: An ESXi host might fail with a purple diagnostic screen due to faulty packets from third-party networking modules
If VMware vSphere Network I/O Control is not enabled, packets from third-party networking modules might cause an ESXi host to fail with a purple diagnostic screen.
This issue is resolved in this release.
- PR 1994373: You might fail to gather data to diagnose networking issues
With ESXi 6.5 Update 2, you can use PacketCapture, a lightweight tcpdump utility implemented in the Reverse HTTP Proxy service, to capture and store only the minimum amount of data to diagnose a networking issue, saving CPU and storage. For more information, see KB 52843.
This issue is resolved in this release.
- PR 1950800: VMkernel adapter fails to communicate through an Emulex physical NIC after an ESXi host boots
After an ESXi host boots, the VMkernel adapter that connects to virtual switches with an Emulex physical NIC might be unable to communicate though the NIC and the VMkernel adapter might not get a DHCP IP address.
This issue is resolved in this release.
- PR 2060589: vSphere Web Client might display incorrect storage allocation after an increase of the disk size of a virtual machine
If you use vSphere Web Client to increase the disk size of a virtual machine, the vSphere Web Client might not fully reflect the reconfiguration and display the storage allocation of the virtual machine as unchanged.
This issue is resolved in this release.
- PR 2012602: An ESXi host might not have access to the Web Services for Management (WSMan) protocol due to failed CIM ticket authentication
An ESXi host might not have access to the WSMan protocol due to failing CIM ticket authentication.
This issue is resolved in this release.
- PR 2065000: Hostd might fail during backup
The ESXi daemon hostd might fail during backup due to a race condition.
This issue is resolved in this release.
- PR 2015082: Deleting files from a content library might cause the vCenter Server Agent to fail
Deleting files from a content library might cause a failure of the vCenter Server Agent (vpxa) service which results in failure of the delete operation.
This issue is resolved in this release.
- PR 1997995: Dell system management software might not display the BIOS attributes of ESXi hosts on Dell OEM servers
Dell system management software might not display the BIOS attributes of ESXi hosts on Dell OEM servers, because the ESXi kernel module fails to load automatically during ESXi boot. This is due to different vendor names in the SMBIOS System Information of Dell and Dell OEM servers.
This issue is resolved in this release.
- PR 1491595: Hostd might fail due to a memory leak
Under certain circumstances, the memory of hostd might exceed the hard limit and the agent might fail.
This issue is resolved in this release.
- PR 2034772: An ESXi host might fail with a purple screen due to a rare race condition in iSCSI sessions
An ESXi host might fail with a purple diagnostic screen due to a rare race condition in iSCSI sessions while processing the scheduled queue and running queue during connection shutdown. The connection lock might be freed to terminate a task while the shutdown is still in process. The next node pointer would be incorrect and cause the failure.
This issue resolved in this release.
- PR 205696: Log congestion leads to unresponsive hosts and inaccessible virtual machines
If a device in a disk group encounters a permanent disk loss, the physical log (PLOG) in the cache device might increase to its maximum threshold. This problem is due to faulty cleanup of data in the log, which are bound to the lost disk. The log entries build up over time, leading to log congestion, which can cause virtual machines to become inaccessible.
This issue is resolved in this release.
- PR 2054940: vSAN health service unable to read controller information
If you are using HP controllers, vSAN health service might be unable to read the controller firmware information. The health check might issue a warning, with the firmware version listed as N/A.
This issue is resolved in this release.
- PR 2034800: Ping failure in vSAN network health check when VMware vRealize Operations manages the vSAN cluster
When vSAN cluster is managed by vRealize Operations, you might observe intermittent ping health check failures.
This issue is resolved in this release.
- PR 2034452: The vSAN health service might be slow displaying data for a large scale cluster after integration with vRealize Operations
After vRealize Operations integration with vSAN Management Pack enabled, the vSAN health service might become less responsive. This problem affects large scale clusters, such as those with 24 hosts, each with 10 disks. vSAN health service uses an inefficient log filter for parsing some specific log patterns, and can consume a large amount of CPU resources while communicating with vRealize Operations.
This issue is resolved in this release.
- PR 2023740: Host might fail with a purple diagnostic screen error if its own local configuration is added through esxcli vsan cluster unicastagent add
When adding unicast configuration of cluster hosts, if the host own local configuration is added, the host experiences a purple diagnostic screen error.
This issue is resolved in this release.
- PR 2033471: Very low IOPS on vSAN iSCSI target when outstanding I/O is 1
For each iSCSI target, the initiator establishes one or more connections depending on whether multipath is configured. For each connection, if there is only one outstanding I/O all the time, and the network is very fast (ping time less than 1 ms), the IOPS on an iSCSI LUN might be much lower that on a native vSAN LUN.
This issue is resolved in this release.
- PR 2028736: A host in a vSAN cluster might fail with a page fault
When the instantiation of a vSAN object fails due to memory constraints, the host might experience a purple diagnostic screen error.
This issue is resolved in this release.
- PR 2023858: vSAN health service alarm could delay 5 to 10 minutes to inform hosts with connection issue
After a host has lost connection with vCenter Server, with no manual retest, there could be a 5 to 10 minute delay for vSAN health service to report the loss of connection.
This issue is resolved in this release.
- PR 2021819: False alarm for cluster configuration consistency health check after enabling vSAN encryption
A false alarm is reported for cluster configuration consistency health check when vSAN encryption is enabled with two or more KMS server configured. The KMS certifications fetched from host and vCenter Server do not match due to improper ordering.
This issue is resolved in this release.
- PR 2059300: Host failure when upgrading to vCenter Server 6.5 Update 2
During upgrade to vCenter Server 6.5 Update 2, when vSAN applies unicast configuration of other hosts in the cluster, a local configuration might be added on the host, resulting in a host failure with purple diagnostic screen.
This issue is resolved in this release.
- PR 2020615: vSphere vMotion operations for virtual machines with more than 64GB of vRAM might fail on vSAN
When you perform a vSphere vMotion operation with a virtual machine with more than 64GB of memory (vRAM) on vSAN, the task might fail due to error in swap file initialization on the destination.
This issue is resolved in this release.
- PR 2008663: After reboot, host is delayed from joining cluster running vSAN 6.6 and later
When there are many vSAN components on disks, during reboot it might take a long time to publish vSAN storage. The host is delayed from joining the vSAN cluster, so this host might be partitioned for a long time.
This issue is resolved in this release.
- PR 2023766: Large amounts of resync traffic for an object being repaired or placed in maintenance mode
When an object spans across all available fault domains, vSAN is unable to replace a component that is impacted by a maintenance mode operation or a repair operation. If this happens, vSAN rebuilds the entire object, which can result in large amounts of resync traffic.
This issue is resolved in this release.
- PR 2039896: Some operations fail and the host must be rebooted
When vSAN Observer is running, memory leaks can occur in the host init group. Other operations that run under the init group might fail until you reboot the host.
This issue is resolved in this release.
- PR 1999439: Slow performance and client latency due to excess DOM I/O activity
Under certain conditions, when I/Os from the Distributed Object Manager (DOM) fail and are retried, the cluster might experience high latencies and slow performance. You might see the following trace logs:
[33675821] [cpu66] [6bea734c OWNER writeWithBlkAttr5 VMDISK] DOMTraceOperationNeedsRetry:3630: {'op': 0x43a67f18c6c0, 'obj': 0x439e90cb0c00, 'objUuid': '79ec2259-fd98-ef85-0e1a-1402ec44eac0', 'status': 'VMK_STORAGE_RETRY_OPERATION'}
This issue is resolved in this release.
- PR 2061175: I/O performance of the input–output memory management unit (IOMMU) driver of AMD might be slow on ESX PCIe devices
I/O performance of ESX PCIe devices might drop when the IOMMU driver of AMD is enabled.
This issue is resolved in this release.
- PR 1982291: Virtual machines with brackets in their names might not deploy successfully
Virtual machines with brackets ([]) in their name might not deploy successfully. For more information, see KB 53648.
This issue is resolved in this release.
-
PR 1950800: VMkernel adapter fails to communicate through an Emulex physical NIC after an ESXi host boots
After an ESXi host boots, the VMkernel adapter that connects to virtual switches with an Emulex physical NIC might be unable to communicate though the NIC and the VMkernel adapter might not get a DHCP IP address.
Workaround: You must unplug the LAN cable and plug it again.
- PR 1998108: An ESXi host might fail with a purple diagnostic screen and a Page Fault error message when you attach a USB device
An ESXi host might fail with a purple diagnostic screen and a
Page Fault
error message when you attach a USB device. This happens because some USB devices report invalid number of interface descriptors to the host.This issue is resolved in this release.
- PR 1948936: Third-party CIM providers or agents might cause memory corruptions that lead to destabilization of the kernel or the drivers of an ESXi host
Third-party CIM providers or agents might cause memory corruptions that lead to destabilization of the kernel or the drivers of an ESXi host and the host might fail with a purple diagnostic screen. Memory corruptions are visible at later point, and not at the exact time of corruption.
This issue is resolved in this release.
- PR 2004950: Update to the VMware Advanced Host Controller Interface driver
The VMware Advanced Host Controller Interface (AHCI) driver is updated to version 1.1.1-1 to fix issues with memory allocation, the Marvell 9230 AHCI controller, and Intel Cannon Lake PCH-H SATA AHCI controller.
- PR 2004948: Update to the nhpsa driver
The nhpsa driver is updated to 2.0.22-1vmw.
- PR 2004949: Updates to the NVMe driver
ESXi 6.5 Update 2 adds management of multiple namespaces compatible with the NVMe 1.2 specification and enhanced diagnostic log.
- PR 2004957: smartpqi driver for HPE Smart PQI controllers
ESXi 6.5 Update 2 enables smartpqi driver support for the HPE ProLiant Gen10 Smart Array Controller.
- PR 2004956: I/O devices enablement for Broadcom SAS 3.5 IT/IR controllers
ESXi 6.5 Update 2 enables a new native driver to support the Broadcom SAS 3.5 IT/IR controllers with devices including combination of NVMe, SAS, and SATA drives. The HBA device IDs are: SAS3516(00AA), SAS3516_1(00AB), SAS3416(00AC), SAS3508(00AD), SAS3508_1(00AE), SAS3408(00AF), SAS3716(00D0), SAS3616(00D1), SAS3708(00D2).
- PR 2004947: Update to the lsi_msgpt3 driver
The lsi_msgpt3 driver is updated to version 16.00.01.00-1vmw.
- PR 2004946: Update to the lsi_mr3 driver to version MR 7.2
The
lsi_mr3
driver is updated to enable support for Broadcom SAS 3.5 RAID controllers. It also contains significant enhancement on task management and critical bug fixes. - PR 2004949: Updates to the NVMe driver
ESXi 6.5 Update 2 adds management of multiple namespaces compatible with the NVMe 1.2 specification and enhanced diagnostic log.
- PR 2015606: VMware ESXi native drivers might fail to support more than eight message signaled interrupts
ESXi native drivers might fail to support more than eight message signaled interrupts even on devices with MSI-X that allows the device to allocate up to 2048 interrupts. This issue is observed on newer AMD-based platforms and is due to default limits of the native drivers, lsi_msgpt2 and lsi_msgpt3.
This issue is resolved in this release.
- PR 2036299: The ESXi native driver for HPE Smart Array controllers extends the list of expander devices to enable compatibility with HPE Gen9 configurations
The disk serviceability plug-in of the ESXi native driver for HPE Smart Array controllers, nhpsa, now works with an extended list of expander devices to enable compatibility with HPE Gen9 configurations.
- PR 2053073: Support for LightPulse and OneConnect adapters is split to separate default drivers
In ESXi 6.5 Update 2, LightPulse and OneConnect adapters are supported by separate default drivers. The brcmfcoe driver supports OneConnect adapters and the lpfc driver supports only LightPulse adapters. Previously, the lpfc driver supported both OneConnect and LightPulse adapters.
- PR 2053703: An ESXi host might fail with purple diagnostic screen when completing installation from a USB media drive on HP servers
When you press Enter to reboot an ESXi host and complete a new installation from a USB media drive on some HP servers, you might see an error on a purple diagnostic screen due to a rare race condition.
This issue is resolved in this release.
- PR 2051280: Some Single Root I/O Virtualization (SR-IOV) graphics configurations that are running ESXi 6.5 with enabled GPU Statistics might cause virtual machines to become unresponsive
Some SR-IOV graphics configurations, that are running ESXi 6.5 with enabled GPU Statistics , might cause virtual machines on the ESXi host to become unresponsive after extended host uptime due to a memory leak. In some cases, the ESXi host might become unresponsive as well.
This issue is resolved in this release.
- The ESXCLI unmap command might reclaim only small file block resources in VMFS-6 datastores and not process large file block resources
In VMFS-6 datastores, if you use the
esxcli storage vmfs unmap
command for manual reclamation of free space, it might only reclaim space from small file blocks and not process large file block resources.This issue is resolved in this release.
- PR 1950800: VMkernel adapter fails to communicate through an Emulex physical NIC after an ESXi host boots
After an ESXi host boots, the VMkernel adapter that connects to virtual switches with an Emulex physical NIC might be unable to communicate though the NIC and the VMkernel adapter might not get a DHCP IP address.
Workaround: You must unplug the LAN cable and plug it again.
- PR 1913483: Third-party storage array user interfaces might not properly display relationships between virtual machine disks and virtual machines
Profile Name | ESXi-6.5.0-20180501001s-standard |
Build | For build information, see Patches Contained in this Release. |
Vendor | VMware, Inc. |
Release Date | May 3, 2018 |
Acceptance Level | PartnerSupported |
Affected Hardware | N/A |
Affected Software | N/A |
Affected VIBs |
|
PRs Fixed | 1827139, 2011288, 1964481 |
Related CVE numbers | N/A |
-
- PR 1827139: ESXi hosts might randomly fail due to errors in sfcbd or other daemon processes
This fix improves internal error handling to prevent ESXi host failures, caused by sfcbd or other daemon processes.
This issue is resolved in this release.
- PR 2011288: Update to glibc package to address security issues
The glibc package has been updated to address CVE-2017-15804, CVE-2017-15670, CVE-2016-4429, CVE-2015-8779, CVE-2015-8778, CVE-2014-9761 and CVE-2015-8776.
- Update of the SQLite database
The SQLite database is updated to version 3.20.1 to resolve a security issue with identifier CVE-2017-10989.
- PR 1827139: ESXi hosts might randomly fail due to errors in sfcbd or other daemon processes
Profile Name | ESXi-6.5.0-20180501001s-no-tools |
Build | For build information, see Patches Contained in this Release. |
Vendor | VMware, Inc. |
Release Date | May 3, 2018 |
Acceptance Level | PartnerSupported |
Affected Hardware | N/A |
Affected Software | N/A |
Affected VIBs |
|
PRs Fixed | 1827139, 2011288, 1964481 |
Related CVE numbers | N/A |
-
- PR 1827139: ESXi hosts might randomly fail due to errors in sfcbd or other daemon processes
This fix improves internal error handling to prevent ESXi host failures, caused by sfcbd or other daemon processes.
This issue is resolved in this release.
- PR 2011288: Update to glibc package to address security issues
The glibc package has been updated to address CVE-2017-15804, CVE-2017-15670, CVE-2016-4429, CVE-2015-8779, CVE-2015-8778, CVE-2014-9761 and CVE-2015-8776.
- Update of the SQLite database
The SQLite database is updated to version 3.20.1 to resolve a security issue with identifier CVE-2017-10989.
- PR 1827139: ESXi hosts might randomly fail due to errors in sfcbd or other daemon processes
Known Issues
The known issues are grouped as follows.
ESXi650-201804201-UG- vSAN Capacity Overview shows incorrect file system overhead
If you have a newly-created vSAN cluster or a cluster storing a small amount of user data, the file system overhead appears to consume a large percentage of the Used Capacity Breakdown. The file system overhead is calculated as a constant, based on the cluster capacity, which is allocated upfront. Since the cluster has very little user data, this constant appears as a large percentage of the used capacity. Once the cluster begins storing user data, the utilization graph becomes more realistic and better reflects the percentages.
Workaround: None.