ESXi 6.7 Update 2 | APR 11 2019 | ISO Build 13006603 |
What's in the Release Notes
The release notes cover the following topics:
- What's New
- Earlier Releases of ESXi 6.7
- Product Support Notices
- Upgrade Notes for This Release
- Patches Contained in this Release
- Resolved Issues
- Known Issues
What's New
- Solarflare native driver: ESXi 6.7 Update 2 adds Solarflare native driver (sfvmk) support for Solarflare 10G and 40G network adaptor devices, such as SFN8542 and SFN8522.
- Virtual Hardware Version 15: ESXi 6.7 Update 2 introduces Virtual Hardware Version 15 which adds support for creating virtual machines with up to 256 virtual CPUs. For more information, see VMware knowledge base articles 1003746 and 2007240.
- Standalone ESXCLI command package: ESXi 6.7 Update 2 provides a new Standalone ESXCLI package for Linux, separate from the vSphere Command Line Interface (vSphere CLI) installation package. The ESXCLI, which is a part of the vSphere CLI, is not updated for ESXi 6.7 Update 2. Although the vSphere CLI installation package is deprecated for this release and is still available for download, you must not install it together with the new Standalone ESXCLI for Linux package. For information about downloading and installing the Standalone ESXCLI package, see VMware {code}.
- In ESXi 6.7 Update 2, the Side-Channel-Aware Scheduler is updated to enhance the compute performance for ESXi hosts that are mitigated for speculative execution hardware vulnerabilities. For more information, see VMware knowledge base article 55806.
- ESXi 6.7 update 2 adds support for VMFS6 automatic unmap processing on storage arrays and devices that report to ESXi hosts an unmap granularity value greater than 1 MB. On arrays that report granularity of 1 MB and less, the unmap operation is supported if the granularity is a factor of 1 MB.
- ESXi 6.7 update 2 adds VMFS6 to the list of supported file systems by the vSphere On-disk Metadata Analyzer (VOMA) to allow you to check and fix issues with VMFS volumes metadata, LVM metadata, and partition table inconsistencies.
Earlier Releases of ESXi 6.7
Features and known issues of ESXi are described in the release notes for each release. Release notes for earlier releases of ESXi 6.7 are:
- VMware ESXi 6.7, Patch Release ESXi670-201903001
- VMware ESXi 6.7, Patch Release ESXi670-201901001
- VMware ESXi 6.7, Patch Release ESXi670-201811001
- VMware ESXi 6.7 Update 1 Release Notes
- VMware vSphere 6.7 Release Notes
For internationalization, compatibility, installation and upgrade, and open source components, see the VMware vSphere 6.7 Release Notes.
Product Support Notices
- VMware vSphere Flash Read Cache is being deprecated. While this feature continues to be supported in the vSphere 6.7 generation, it will be discontinued in a future vSphere release. As an alternative, you can use the vSAN caching mechanism or any VMware certified third-party I/O acceleration software listed in the VMware Compatibility Guide.
Upgrade Notes for This Release
For more information on ESXi versions that support upgrade to ESXi 6.7 Update 2, please see VMware knowledge base article 67077.
Patches Contained in this Release
This release contains all bulletins for ESXi that were released earlier to the release date of this product.
Build Details
Download Filename: | update-from-esxi6.7-6.7_update02.zip |
Build: | 13006603 12986307 (Security-only) |
Download Size: | 453.0 MB |
md5sum: |
afb6642c5f78212683d9d9a1e9dde6bc |
sha1checksum: |
783fc0a2a457677e69758969dadc6421e5a7c17a |
Host Reboot Required: | Yes |
Virtual Machine Migration or Shutdown Required: | Yes |
Bulletins
This release contains general and security-only bulletins. Security-only bulletins are applicable to new security fixes only. No new bug fixes are included, but bug fixes from earlier patch and update releases are included.
If the installation of all new security and bug fixes is required, you must apply all bulletins in this release. In some cases, the general release bulletin will supersede the security-only bulletin. This is not an issue as the general release bulletin contains both the new security and bug fixes.
The security-only bulletins are identified by bulletin IDs that end in "SG". For information on patch and update classification, see KB 2014447.
For more information about the individual bulletins, see the My VMware page and the Resolved Issues section.
Bulletin ID | Category | Severity |
ESXi670-201904201-UG | Bugfix | Critical |
ESXi670-201904202-UG | Bugfix | Important |
ESXi670-201904203-UG | Bugfix | Important |
ESXi670-201904204-UG | Bugfix | Important |
ESXi670-201904205-UG | Bugfix | Important |
ESXi670-201904206-UG | Bugfix | Important |
ESXi670-201904207-UG | Bugfix | Moderate |
ESXi670-201904208-UG | Bugfix | Important |
ESXi670-201904209-UG | Enhancement | Important |
ESXi670-201904210-UG | Enhancement | Important |
ESXi670-201904211-UG | Bugfix | Critical |
ESXi670-201904212-UG | Bugfix | Important |
ESXi670-201904213-UG | Bugfix | Important |
ESXi670-201904214-UG | Bugfix | Important |
ESXi670-201904215-UG | Bugfix | Moderate |
ESXi670-201904216-UG | Bugfix | Important |
ESXi670-201904217-UG | Bugfix | Important |
ESXi670-201904218-UG | Bugfix | Important |
ESXi670-201904219-UG | Bugfix | Important |
ESXi670-201904220-UG | Bugfix | Important |
ESXi670-201904221-UG | Bugfix | Important |
ESXi670-201904222-UG | Bugfix | Moderate |
ESXi670-201904223-UG | Bugfix | Important |
ESXi670-201904224-UG | Bugfix | Important |
ESXi670-201904225-UG | Enhancement | Important |
ESXi670-201904226-UG | Bugfix | Important |
ESXi670-201904227-UG | Bugfix | Important |
ESXi670-201904228-UG | Bugfix | Important |
ESXi670-201904229-UG | Bugfix | Important |
ESXi670-201904101-SG | Security | Important |
ESXi670-201904102-SG | Security | Important |
ESXi670-201904103-SG | Security | Important |
IMPORTANT: For clusters using VMware vSAN, you must first upgrade the vCenter Server system. Upgrading only ESXi is not supported.
Before an upgrade, always verify in the VMware Product Interoperability Matrix compatible upgrade paths from earlier versions of ESXi and vCenter Server to the current version.
Image Profiles
VMware patch and update releases contain general and critical image profiles.
Application of the general release image profile applies to new bug fixes.
Image Profile Name |
ESXi-6.7.0-20190402001-standard |
ESXi-6.7.0-20190402001-no-tools |
ESXi-6.7.0-20190401001s-standard |
ESXi-6.7.0-20190401001s-no-tools |
Patch Download and Installation
The typical way to apply patches to ESXi hosts is through the VMware vSphere Update Manager. For details, see the About Installing and Administering VMware vSphere Update Manager.
ESXi hosts can be updated by manually downloading the patch ZIP file from the VMware download page and installing the VIB by using the esxcli software vib command. Additionally, the system can be updated using the image profile and the esxcli software profile command.
For more information, see the vSphere Command-Line Interface Concepts and Examples and the vSphere Upgrade Guide.
Resolved Issues
The resolved issues are grouped as follows.
- ESXi670-201904201-UG
- ESXi670-201904202-UG
- ESXi670-201904203-UG
- ESXi670-201904204-UG
- ESXi670-201904205-UG
- ESXi670-201904206-UG
- ESXi670-201904207-UG
- ESXi670-201904208-UG
- ESXi670-201904209-UG
- ESXi670-201904210-UG
- ESXi670-201904211-UG
- ESXi670-201904212-UG
- ESXi670-201904213-UG
- ESXi670-201904214-UG
- ESXi670-201904215-UG
- ESXi670-201904216-UG
- ESXi670-201904217-UG
- ESXi670-201904218-UG
- ESXi670-201904219-UG
- ESXi670-201904220-UG
- ESXi670-201904221-UG
- ESXi670-201904222-UG
- ESXi670-201904223-UG
- ESXi670-201904224-UG
- ESXi670-201904225-UG
- ESXi670-201904226-UG
- ESXi670-201904227-UG
- ESXi670-201904228-UG
- ESXi670-201904229-UG
- ESXi670-201904101-SG
- ESXi670-201904102-SG
- ESXi670-201904103-SG
- ESXi-6.7.0-20190402001-standard
- ESXi-6.7.0-20190402001-no-tools
- ESXi-6.7.0-20190401001s-standard
- ESXi-6.7.0-20190401001s-no-tools
Patch Category | Bugfix |
Patch Severity | Critical |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2227123, 2250472, 2273186, 2282052, 2267506, 2256771, 2269616, 2250697, 2193516, 2203431, 2264996, 2267158, 2231707, 2191342, 2197795, 2221935, 2193830, 2210364, 2177107, 2179263, 2185906, 2204508, 2211350, 2204204, 2226565, 2240385, 2163397, 2238134, 2213883, 2141221, 2240444, 2249713, 2184068, 2257480, 2235031, 2257354, 2212140, 2279897, 2227623, 2277648, 2221778, 2287560, 2268541, 2203837, 2236531, 2242656, 2250653, 2260748, 2256001, 2266762, 2263849, 2267698, 2271653, 2232538, 2210056, 2271322, 2272737, 2286563, 2178119, 2223879, 2221256, 2258272, 2219661, 2268826, 2224767, 2257590, 2241917, 2246891, 2269308 |
CVE numbers | N/A |
This patch updates the esx-base, esx-update, vsan
and vsanhealth
VIBs to resolve the following issues:
- PR 2227123: Very large values of counters for read or write disk latency might trigger alarms
You might intermittently see alarms for very large values of counters for read or write disk latency, such as
datastore.totalReadLatency
anddatastore.totalWriteLatency
.This issue is resolved in this release.
- PR 2250472: Linux virtual machines might stop responding while powering on due to some memory size configurations
When a Linux virtual machine is configured with specific memory size, such as 2052 MB or 2060 MB, instead of powering on, it might display a blank screen.
This issue is resolved in this release.
- PR 2273186: An ESXi host might fail when you enable a vSphere Distributed Switch health check
When you set up the network and enable vSphere Distributed Switch health check to perform configuration checks, the ESXi host might fail with purple diagnostic screen.
This issue is resolved in this release.
- PR 2282052: The SNMP agent might deliver an incorrect trap vmwVmPoweredOn
The SNMP agent might deliver an incorrect trap
vmwVmPoweredOn
when a virtual machine is selected under the Summary tab and the Snapshot Manager tab in the vSphere Web Client.This issue is resolved in this release.
- PR 2267506: An ESXi host might fail with purple diagnostic screen at DVFilter level
The DVFilter might receive unexpected or corrupt values in the shared memory ring buffer, causing the internal function to return
NULL
. If thisNULL
value is not handled gracefully, an ESXi host might fail with purple diagnostic screen at DVFilter level.This issue is resolved in this release.
- PR 2256771: You might see an unnecessary event message for network connectivity restored on virtual switch when powering on or off virtual machines
In the Events tab of the vSphere Client or the vSphere Web Client, when you power on or off virtual machines, you might see events similar to
Network connectivity restored on virtual switch XXX, portgroups: XXX. Physical NIC vmnicX is up
. When checking uplink status, the system report might include port groups that are not affected by the virtual machine powering on or off operation, which triggers the event. However, the event does not indicate a problem or prompt an action.This issue is resolved in this release. The event message is removed.
- PR 2250697: Windows Server Failover Cluster validation might fail if you configure Virtual Volumes with a Round Robin path policy
If during the Windows Server Failover Cluster setup you change the default path policy from Fixed or Most Recently Used to Round Robin, the I/O of the cluster might fail and the cluster might stop responding.
This issue is resolved in this release.
- PR 2193516: Host Profile remediation might end with deleting a virtual network adapter from the VMware vSphere Distributed Switch
If you extract a Host Profile from an ESXi host with disabled IPv6 support, and the IPv4 default gateway is overridden while remediating that Host Profile, you might see a message that a virtual network adapter, such as DSwitch0, is to be created on the vSphere Distributed Switch (VDS), but the adapter is actually removed from the VDS. This is due to a "::" value of the
ipV6DefaultGateway
parameter.This issue is resolved in this release.
- PR 2203431: Some guest virtual machines might report their identity as an empty string
Some guest virtual machines might report their identity as an empty string if the guest is running a later version of VMware Tools than the ESXi host was released with, or if the host cannot recognize the
guestID
parameter of theGuestInfo
object. The issue affects mainly virtual machines with CentOS and VMware Photon OS.This issue is resolved in this release.
- PR 2264996: You see repetitive messages in the vCenter Web Client prompting you to delete host sub specifications nsx
You might see repetitive messages in the vCenter Web Client, similar to
Delete host sub specification nsx
if you use network booting for ESXi hosts, due to some dummy code in hostd.This issue is resolved in this release.
- PR 2231707: The virtual machine might stop responding if the AutoInstalls ISO image file is mounted
The installation of the VMware Tools unmounts the ISO image file that ends with
autoinst.iso
, if it is already mounted to the virtual machine. To mount the VMware Tools ISO image file, you must click Install VMware Tools again. As a result, the virtual machine might stop responding.This issue is resolved in this release.
- Pr 2267158: ESXi hosts might lose connectivity to the vCenter Sever system due to failure of the vpxa agent service
ESXi hosts might lose connectivity to the vCenter Sever system due to failure of the vpxa agent service in result of an invalid property update generated by the
PropertyCollector
object. A race condition in hostd leads to a malformed sequence of property update notification events that cause the invalid property update.This issue is resolved in this release.
- PR 2191342: The SMART disk monitoring daemon, smartd, might flood the syslog service logs of release builds with debugging and info messages
In release builds, the smartd daemon might generate a lot of debugging and info messages into the syslog service logs.
This issue is resolved in this release. The fix removes debug messages from release builds.
- PR 2197795: The scheduler of the VMware vSphere Network I/O Control (NIOC) might intermittently reset the uplink network device
The NIOC scheduler might reset the uplink network device if the uplink is rarely used. The reset is non-predictable.
This issue is resolved in this release.
- PR 2221935: ESXi hosts might intermittently fail if you disable global IPv6 addresses
ESXi hosts might intermittently fail if you disable global IPv6 addresses, because a code path still uses IPv6.
This issue is resolved in this release. If you already face the issue, re-enable the global IPv6 addresses to avoid failure of hosts. If you need to disable IPv6 addresses for some reason, you must disable IPv6 on individual vmknics, not global IPv6 addresses.
- PR 2193830: Тhe virtual machine executable process might fail and shut down virtual machines during a reboot of the guest OS
Due to a race condition, the virtual machine executable process might fail and shut down virtual machines during a reboot of the guest OS.
This issue is resolved in this release.
- PR 2210364: The ESXi daemon hostd might become unresponsive while waiting for an external process to finish
Hostd might become unresponsive while waiting for an external process initiated with the
esxcfg-syslog
command to finish.This issue is resolved in this release.
- PR 2177107: A soft lockup of physical CPUs might cause an ESXi host to fail with a purple diagnostic screen
Large number of I/Os getting timeout on a heavy-loaded system might cause a soft lockup of physical CPUs. As a result, an ESXi host might fail with a purple diagnostic screen.
This issue is resolved in this release.
- PR 2179263: Virtual machines with a virtual SATA CD-ROM might fail due to invalid commands
Invalid commands to a virtual SATA CD-ROM might trigger errors and increase the memory usage of virtual machines. This might lead to a failure of virtual machines if they are unable to allocate memory. You might see the following logs for invalid commands:
YYYY-MM-DDTHH:MM:SS.mmmZ| vcpu-1| I125: AHCI-USER: Invalid size in command
YYYY-MM-DDTHH:MM:SS.mmmZ| vcpu-1| I125: AHCI-VMM: sata0:0: Halted the port due to command abort.And a similar panic message:
YYYY-MM-DDTHH:MM:SS.mmmZ| vcpu-1| E105: PANIC: Unrecoverable memory allocation failure
This issue is resolved in this release.
- PR 2185906: Migration of virtual machines by using VMware vSphere vMotion might fail with a NamespaceDb compatibility error if Guest Introspection service is on
If the Guest Introspection service in a vSphere 6.7 environment with more than 150 virtual machines is active, migration of virtual machines by using vSphere vMotion might fail with an error in the vSphere Web Client similar to
The source detected that the destination failed to resume
.
The destinationvmware.log
contains error messages similar to:2018-07-18T02:41:32.035Z| vmx| I125: MigrateSetState: Transitioning from state 11 to 12.
2018-07-18T02:41:32.035Z| vmx| I125: Migrate: Caching migration error message list:
2018-07-18T02:41:32.035Z| vmx| I125: [msg.checkpoint.migration.failedReceive] Failed to receive migration.
2018-07-18T02:41:32.035Z| vmx| I125: [msg.namespaceDb.badVersion] Incompatible version -1 (expect 2).
2018-07-18T02:41:32.035Z| vmx| I125: [msg.checkpoint.mrestoregroup.failed] An error occurred restoring the virtual machine state during migration.The vmkernel log contains error messages such as:
2018-07-18T02:32:43.011Z cpu5:66134)WARNING: Heap: 3534: Heap fcntlHeap-1 already at its maximum size. Cannot expand.
2018-07-18T02:41:35.613Z cpu2:66134)WARNING: Heap: 4169: Heap_Align(fcntlHeap-1, 200/200 bytes, 8 align) failed. caller: 0x41800aaca9a3This issue is resolved in this release.
- PR 2204508: The advanced performance chart might stop displaying data after a hostd restart
You can edit the statistical data to be collected only when the virtual machine is in a particular power state. For example, you cannot edit the memory data when the virtual machine is powered on. When you restart the hostd service, the virtual machine might not change its power state, which leads to false data initialization. The calculation of the host consumed memory counter might end up with division by zero and the chart might stop drawing graphs.
This issue is resolved in this release.
- PR 2211350: A virtual machine port might be blocked after a VMware vSphere High Availability (vSphere HA) failover
By default, the Isolation Response feature in an ESXi host with enabled vSphere HA is disabled. When the Isolation Response feature is enabled, a virtual machine port, which is connected to an NSX-T logical switch, might be blocked after a vSphere HA failover.
This issue is resolved in this release.
- PR 2204204: e1000 vNICs might intermittently drop unrecognized GRE packets
e1000 vNICs might intermittently drop unrecognized GRE packets due to a parsing issue.
This issue is resolved in this release.
- PR 2226565: Booting ESXi on Apple hardware might fail with an error that the Mutiboot buffer is too small
While booting an Apple system, the boot process might stop with the following error message:
Shutting down firmware services...
Mutiboot buffer is too small.
Unrecoverable errorThe issue affects only certain Apple firmware versions.
This issue is resolved in this release.
- PR 2240385: Virtual machines might fail during a long-running quiesced snapshot operation
If hostd restarts during a long-running quiesced snapshot operation, hostd might automatically run the snapshot Consolidation command to remove redundant disks and improve virtual machine performance. However, the Consolidation command might race with the running quiesced snapshot operation and cause failure of virtual machines.
This issue is resolved in this release.
- PR 2163397: The SNMP agent might display D_Failed battery status
The SNMP agent might display the status of all batteries as D_Failed when you run the
snmpwalk
command. This issue occurs due to the processing of code, which might not check the status properly and the misinterpretation of the compact sensors with sensor-specific code.This issue is resolved in this release.
- PR 2238134: When you use an IP hash or source MAC hash teaming policy, the packets might drop or go through the wrong uplink
When you use an IP hash or source MAC hash teaming policy, some packets from the packet list might use a different uplink. As a result, some of them might drop or not be sent out through the uplink determined by the teaming policy.
This issue is resolved in this release.
- PR 2213883: An ESXi host becomes unresponsive when it disconnects from an NFS datastore that is configured for logging
If an NFS datastore is configured as a syslog datastore and if the ESXi host disconnects from it, logging to the datastore stops and the ESXi host might become unresponsive.
This issue is resolved in this release.
- PR 2141221: An ESXi host might fail during replication of virtual machines by using VMware vSphere Replication with VMware Site Recovery Manager
When you replicate virtual machines by using vSphere Replication with Site Recovery Manager, the ESXi host might fail with a purple diagnostic screen immediately or within 24 hours. You might see a similar error:
PANIC bora/vmkernel/main/dlmalloc.c:4924 - Usage error in dlmalloc
.This issue is resolved in this release.
- PR 2240444: The hostd daemon might fail with an error due to too many locks
Larger configurations might exceed the limit for the number of locks and hostd starts failing with an error similar to:
hostd Panic: MXUserAllocSerialNumber: too many locks!
This issue is resolved in this release. The fix removes the limit for the number of locks.
- PR 2249713: The esxcli network plug-in command might return a false value as a result
Тhe esxcli
$esxcli.network.nic.coalesce.set.Invoke
network command might return a false value as a result even though the value is correctly set. This might impact your automation scripts.This issue is resolved in this release.
- PR 2184068: An ESXi host might fail with a purple diagnostic screen when accessing a corrupted VMFS datastore
When a virtual machine tries to access a VMFS datastore with corrupted directory, the ESXi host might fail with a purple diagnostic screen and the
DirEntry corruption
error message.This issue is resolved in this release.
- PR 2257480: The LACPPDU Actor кey value might change to 0
The LACPPDU Actor key might change its value to 0 during the link status flapping.
This issue is resolved in this release.
- PR 2235031: An ESXi host becomes unresponsive and you see warnings for reached maximum heap size in the vmkernel.log
Due to a timing issue in the VMkernel, buffers might not be flushed, and the heap gets exhausted. As a result, services such as hostd, vpxa and vmsyslogd might not be able to write logs on the ESXi host, and the host becomes unresponsive. In the
/var/log/vmkernel.log
, you might see a similar warning:WARNING: Heap: 3571: Heap vfat already at its maximum size. Cannot expand.
This issue is resolved in this release.
- PR 2257354: Some virtual machines might become invalid when trying to install or update VMware Tools
If you have modified the type of a virtual CD-ROM device of a virtual machine in a previous version of ESXi, after you update to a later version, when you try to install or update VMware Tools, the virtual machine might be terminated and marked as invalid.
This issue is resolved in this release.
- PR 2212140: Renewing a host certificate might not push the full chain of trust to the ESXi host
When you renew a certificate, only the first certificate from the supplied chain of trust might be stored on the ESXi host. Any intermediate CA certificates are truncated. Because of the missing certificates, the chain to the root CA cannot be built. This leads to a warning for an untrusted connection.
This issue is resolved in this release.
- PR 2279897: Creating a snapshot of a virtual machine might fail due to a null VvolId parameter
If a vSphere API for Storage Awareness provider modifies the vSphere Virtual Volumes policy unattended, a null
VvolID
parameter might update the vSphere Virtual Volumes metadata. This results in a VASA call with a nullVvoId
parameter and a failure when creating a virtual machine snapshot.This issue is resolved in this release. The fix handles the policy modification failure and prevents the null
VvolId
parameter. - PR 2227623: Parallel cloning of multiple virtual machines on a vSphere Virtual Volumes datastore might fail with an error message for failed file creation
If a call from a vSphere API for Storage Awareness provider fails due to all connections to the virtual provider being busy, operations for parallel cloning of multiple virtual machines on a vSphere Virtual Volumes datastore might become unresponsive or fail with an error message similar to
Cannot complete file creation operation
.This issue is resolved in this release.
- PR 2277648: An ESXi host might fail with a purple diagnostic screen displaying a sbflush_internal panic message
An ESXi host might fail with a purple diagnostic screen displaying a
sbflush_internal
panic message due to some discrepancies in the internal statistics.This issue is resolved in this release. The fix converts the panic message into an assert.
- PR 2221778: You cannnot migrate virtual machines by using vSphere vMotion between ESXi hosts with NSX managed virtual distributed switches (N-VDS) and vSphere Standard Switches
With ESXi 6.7 Update 2, you can migrate virtual machines by using vSphere vMotion between ESXi hosts with N-VDS and vSphere Standard Switches. To enable the feature, you must upgrade your vCenter Server system to vCenter Server 6.7 Update 2 and ESXi 6.7 Update 2 on both source and destination sites.
This issue is resolved in this release.
- PR 2287560: ESXi hosts might fail with a purple diagnostic screen due to a page fault exception
ESXi hosts might fail with a purple diagnostic screen due to a page fault exception. The error comes from a missing flag that could trigger cache eviction on an object that no longer exists, and result in a
NULL
pointer reference.This issue is resolved in this release.
- PR 2268541: You might see redundant VOB messages in the vCenter Server system, such as Lost uplink redundancy on virtual switch "xxx"
When checking the uplink status, the teaming policy might not check if the uplink belongs to a port group and the affected port group might not be correct. As a result, you might see redundant VOB messages in the
/var/run/log/vobd.log
file, such as:Lost uplink redundancy on virtual switch "xxx". Physical NIC vmnicX is down. Affected portgroups:"xxx"
.This issue is resolved in this release.
- PR 2203837: Claim rules must be manually added to an ESXi host
You must manually add the claim rules to an ESXi host for Lenovo ThinkSystem DE Series Storage Arrays.
This issue is resolved in this release. This fix sets Storage Array Type Plugin (SATP) to
VMW_SATP_ALUA
, Path Selection Policy (PSP) toVMW_PSP_RR
, and Claim Options totpgs_on
as default for Lenovo ThinkSystem DE Series Storage Arrays. - PR 2236531: An ESXi host might get disconnected from a vCenter Server system
The vpxa service might fail due to excessive memory usage when iterating on a large directory tree. As a result, an ESXi host might get disconnected from a vCenter Server system.
This issue is resolved in this release.
- PR 2242656: Updates with ESXCLI commands might fail when SecureBoot is enabled
The commands
esxcli software profile update
oresxcli software profile install
might fail on hosts running ESXi 6.7 Update 1 with enabled SecureBoot. You might see error messages such asFailed to setup upgrade using esx-update VIB
andFailed to mount tardisk in ramdisk: [Errno 1] Operation not permitted
.This issue is resolved in this release. If you already face the issue, run the
esxcli software vib update
command to update your system to ESXi 6.7 Update 2. You can also disable SecureBoot, run the ESXCLI commands, and then re-enable SecureBoot. - PR 2250653: NFS volume mount might not persist after reboot of an ESXi host
NFS volume mount might not persist after a reboot of an ESXi host due to intermittent failure in resolving the host name of the NFS server.
This issue is resolved in this release. The fix adds retry logic if a host name resolution fails.
- PR 2260748: You might intermittently lose connectivity while using the Route Based on Physical NIC Load option in the vSphere Distributed Switch
This issue happens if you create a VMK or vNIC port by using the Route Based on Physical NIC Load option when the physical NIC assigned to that port group is temporarily down. In such cases, the port team uplink is not set because there are no active uplinks in the portgroup. When the physical NIC becomes active again, teaming code fails to update the port data, resulting in a packet loss.
This issue is resolved in this release.
- PR 2256001: Host fails with Bucketlist_LowerBound or PLOGRelogRetireLsns in the backtrace
A race condition can occur when Plog Relog runs along with DecommissionMD, causing Relog to access freed Plog device state tables. This problem can cause a host failure with a purple diagnostic screen.
This issue is resolved in this release.
- PR 2266762: You might see many warnings for unsupported physical block sizes in the vmkernel logs
If you use devices with physical block size other than 4096 or 512 MB, in the vmkernel logs you might see many warnings similar to:
ScsiPath: 4395: The Physical block size "8192" reported by the path vmhba3:C0:T1:L0 is not supported.
This issue is resolved in this release. The fix converts the warnings into debug log messages.
- PR 2263849: Operations on stretched cluster failed with name '_VSAN_VER3' is not defined
After configuring or upgrading a vSAN stretched cluster, all cluster level requests might fail. You might see the following error message:
name '_VSAN_VER3' is not defined
.This issue is resolved in this release.
- PR 2267698: vSAN performance charts do not show VMDK metrics properly
If the virtual machine folder is created using a name instead of UUID, the vSAN performance charts do not show the following VMDK metrics properly:
- IOPS and IOPS Limit
- Delayed Normalized IOPSThis issue is resolved in this release.
- PR 2271653: You might see wrong reports for dropped packets when using IOChain and the Large Receive Offload (LRO) feature is enabled
In Host > Monitor > Performance > Physical Adapters > pNIC vSwitch Port Drop Rate you might see a counter showing dropped packets, while no packets are dropped. This is because the IOChain framework treats the reduced number of packets, when using LRO, as packet drops.
This issue is resolved in this release.
- PR 2232538: An ESXi host might fail while powering on a virtual machine with a name that starts with “SRIOV”
An ESXi host might fail with a purple diagnostic screen while powering on a virtual machine. This happens if the virtual machine name starts with “SRIOV” and has an uplink connected to a vSphere Distributed Switch.
This issue is resolved in this release.
- PR 2210056: If SNMP disables dynamic rulesets, but they are active in a Host Profile, you might see an unexpected compliance error
In a rare race condition, when SNMP disables dynamic rulesets, but they are active in a Host Profile, you might see an unexpected compliance error such as
Ruleset dynamicruleset not found
.This issue is resolved in this release.
- PR 2271322: Recovery of a virtual machine running Windows 2008 or later by using vSphere Replication with application quiescing enabled might result in a corrupt replica
Recovery of a virtual machine running Windows 2008 or later by using vSphere Replication with application quiescing enabled might result in a corrupt replica. Corruption only happens when application quiescing is enabled. If VMware Tools on the virtual machine is not configured for application quiescing, vSphere Replication uses a lower level of consistency, such as file system quiescing, and replicas are not corrupted.
This issue is resolved in this release.
- PR 2272737: When you perform CIM requests more than 2 MB, the SFCB configuration might fail on an ESXi host
When you perform CIM requests more than 2 MB like firmware download, CIM service might fail on an ESXi host due to out of stack space and the
SFCB zdumb
file appears in/var/core
.This issue is resolved in this release.
- PR 2286563: The number of syslog.log archive files generated might be less than the configured default and different between ESXi versions 6.0 and 6.7
The number of
syslog.log
archive files generated might be one less than the default value set by using thesyslog.global.defaultRotate
parameter. The number ofsyslog.log
archive files might also be different between ESXi versions 6.0 and 6.7. For instance, ifsyslog.global.defaultRotate
is set to 8 by default, ESXi 6.0 createssyslog.0.gz
tosyslog.7.gz
, while ESXi 6.7 createssyslog.0.gz
tosyslog.6.gz
.This issue is resolved in this release.
- PR 2178119: A large value of the advanced config option /Misc/MCEMonitorInterval might cause ESXi hosts to become unresponsive
If you set the advanced config option
/Misc/MCEMonitorInterval
to a very large value, ESXi hosts might become unresponsive in around 10 days of uptime, due to a timer overflow.This issue is resolved in this release.
- PR 2223879: An ESXi host might fail with a purple diagnostic screen and an error for a possible deadlock
The P2MCache lock is a spin lock causing slow CPU performance that might cause starvation of other CPUs. As a result, an ESXi host might fail with a purple diagnostic screen and a
Spin count exceeded - possible deadlock, PVSCSIDoIOComplet, NFSAsyncIOComplete
error.This issue is resolved in this release.
- PR 2221256: An ESXi host might fail with a purple diagnostic screen during a Storage vMotion operation between NFS datastores
An ESXi host might fail with a purple diagnostic screen during a Storage vMotion operation between NFS datastores, due to a race condition in the request completion path. The race condition happens under heavy I/O load in the NFS stack.
This issue is resolved in this release.
- PR 2258272: ESXi host with vSAN enabled fails due to race condition in decommission workflow
During disk decommission operation, a race condition in the decommission workflow might cause the ESXi host to fail with a purple diagnostic screen.
This issue is resolved in this release.
- PR 2219661: An ESXi host might fail with a purple diagnostic screen if RSS is enabled on a physical NIC
An ESXi host might fail with a purple diagnostic screen if RSS is enabled on a physical NIC. You might see a similar backtrace:
yyyy-mm-dd cpu9:2097316)0x451a8521bc70:[0x41801a68ad30]RSSPlugCleanupRSSEngine@(lb_netqueue_bal)#
+0x7d stack: 0x4305e4c1a8e0, 0x41801a68af2b, 0x430db8c6b188, 0x4305e4c1a8e0,
0x0yyyy-mm-dd cpu9:2097316)0x451a8521bc90:[0x41801a68af2a]RSSPlugInitRSSEngine@(lb_netqueue_bal)#+0x127 stack: 0x0, 0x20c49ba5e353f7cf, 0x4305e4c1a930, 0x4305e4c1a800, 0x4305e4c1a9f0
yyyy-mm-dd cpu9:2097316)0x451a8521bcd0:[0x41801a68b21c]RSSPlug_PreBalanceWork@(lb_netqueue_bal)#+0x1cd stack: 0x32, 0x32, 0x0, 0xe0, 0x4305e4c1a800
yyyy-mm-dd cpu9:2097316)0x451a8521bd30:[0x41801a687752]Lb_PreBalanceWork@(lb_netqueue_bal)#+0x21f stack: 0x4305e4c1a800, 0xff, 0x0, 0x43057f38ea00, 0x4305e4c1a800
yyyy-mm-dd cpu9:2097316)0x451a8521bd80:[0x418019c13b18]UplinkNetqueueBal_BalanceCB@vmkernel#nover+0x6f1 stack: 0x4305e4bc6088, 0x4305e4c1a840, 0x4305e4c1a800, 0x4305e435fb30, 0x4305e4bc6088
yyyy-mm-dd cpu9:2097316)0x451a8521bf00:[0x418019cd1c89]UplinkAsyncProcessCallsHelperCB@vmkernel#nover+0x116 stack: 0x430749e60870, 0x430196737070, 0x451a85223000, 0x418019ae832b, 0x430749e60088
yyyy-mm-dd cpu9:2097316)0x451a8521bf30:[0x418019ae832a]HelperQueueFunc@vmkernel#nover+0x157 stack: 0x430749e600b8, 0x430749e600a8, 0x430749e600e0, 0x451a85223000, 0x430749e600b8
yyyy-mm-dd cpu9:2097316)0x451a8521bfe0:[0x418019d081f2]CpuSched_StartWorld@vmkernel#nover+0x77 stack: 0x0, 0x0, 0x0, 0x0, 0x0This issue is resolved in this release. The fix cleans data in the RSS engine to avoid overloading of the dedicated heap of the NetQueue balancer.
- PR 2268826: An ESXi host might fail with a purple diagnostic screen when the VMware APIs for Storage Awareness (VASA) provider sends a rebind request to switch the protocol endpoint for a vSphere Virtual Volume
When the VASA provider sends a rebind request to an ESXi host to switch the binding for a particular vSphere Virtual Volume driver, the ESXi host might switch the protocol endpoint and other resources to change the binding without any I/O disturbance. As a result, the ESXi host might fail with a purple diagnostic screen.
This issue is resolved in this release.
- PR 2224767: You might see false hardware health alarms from Intelligent Platform Management Interface (IPMI) sensors
Some IPMI sensors might intermittently change from green to red and the reverse, which creates false hardware health alarms.
This issue is resolved in this release. With this fix, you can use the advanced option command
esxcfg-advcfg -s {sensor ID},{sensor ID} /UserVars/HardwareHealthIgnoredSensors
to ignore hardware health alarms from selected sensors. - PR 2257590: An ESXi host might fail with a memory admission error if you install an async 32-bit iSCSI Management API (IMA) plug-in
If you install an async 32-bit IMA plug-in and there are already installed 32-bit IMA plug-ins, the system might try to load these plug-ins and check if they are loaded from the QLogic independent iSCSI adapter. During this process the ESXi host might fail with a memory admission error.
This issue is resolved in this release. This fix loads each 32-bit IMA plug-in only from the QLogic adapter.
- PR 2241917: A reprotect operation might fail with an error for vSphere Replication synchronization
If you run a reprotect operation by using the Site Recovery Manager, the operation might fail if a virtual machine is late with the initial synchronization. You might see an error similar to:
VR synchronization failed for VRM group {…}. Operation timed out: 10800 seconds
This issue is resolved in this release.
- PR 2246891: An I/O error TASK_SET_FULL on a secondary LUN might slow down the I/O rate on all secondary LUNs behind the protocol endpoint of Virtual Volumes on HPE 3PAR storage if I/O throttling is enabled
When I/O throttling is enabled on a protocol endpoint of Virtual Volumes on HPE 3PAR storage and if an I/O on a secondary LUN fails with an error
TASK_SET_FULL
, the I/O rate on all secondary LUNs that are associated with the protocol endpoint slows down.This issue is resolved in this release. With this fix, you can enable I/O throttling on individual Virtual Volumes to avoid the slowdown of all secondary LUNs behind the protocol endpoint if the
TASK_SET_FULL
error occurs. - PR 2269308: ESXi hosts with visibility to RDM LUNs might take a long time to start or during LUN rescan
The large number of RDM LUNs might cause an ESXi host to take a long time to start or delay while performing a LUN rescan. If you use APIs, such as
MarkPerenniallyReserved
orMarkPerenniallyReservedEx
, you can mark a specific LUN as perennially reserved, which improves the start and time and rescan time of the ESXi hosts.This issue is resolved in this release.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the i40en
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the igbn
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the ixgben
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the ne1000
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the vmkusb
VIB.
Patch Category | Bugfix |
Patch Severity | Moderate |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2228529 |
CVE numbers | N/A |
This patch updates the qlnativefc
VIB.
- Update to the qlnativefc driver
The qlnativefc driver is updated to version 3.1.8.0.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2231435 |
CVE numbers | N/A |
This patch updates the nfnic
VIB.
- Update to the nfnic driver
The nfnic driver is updated to version 4.0.0.17.
Patch Category | Bugfix |
Patch Severity | Enhancement |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the brcmfcoe
VIB.
Patch Category | Bugfix |
Patch Severity | Enhancement |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included | VMW_bootbank_lpfc_11.4.33.18-12vmw.670.2.48.13006603 |
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the lpfc
VIB.
Patch Category | Bugfix |
Patch Severity | Critical |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2226688 |
CVE numbers | N/A |
This patch updates the elx-esx-libelxima.so
VIB to resolve the following issue:
- PR 2226688: Emulex drivers logs might fill up the /var file system logs
Emulex drivers might write logs at
/var/log/EMU/mili/mili2d.log
and fill up the 40 MB/var
file system logs of RAM drives.This issue is resolved in this release. The fix changes writes of Emulex drivers to the
/scratch/log/
instead of the/var/log/
.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the misc-drivers
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the lsu-lsi-lsi-msgpt3-plugin
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | No |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2234599 |
CVE numbers | N/A |
This patch updates the sfvmk
VIB.
ESXi 6.7 Update 2 adds Solarflare native driver (sfvmk) support for Solarflare 10G and 40G network adaptor devices, such as SFN8542 and SFN8522.
Patch Category | Bugfix |
Patch Severity | Moderate |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2256555 |
CVE numbers | N/A |
This patch updates the lsi-msgpt2
VIB.
- Update to the lsi-msgpt2 driver
The lsi-msgpt2 driver is updated to version 20.00.05.00 to address task management enhancements and critical bug fixes.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the nmlx4-core
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the nmlx4-en
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the nmlx4-rdma
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the nmlx5-core
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the nmlx5-rdma
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the bnxtnet
VIB.
Patch Category | Bugfix |
Patch Severity | Moderate |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2242441 |
CVE numbers | N/A |
This patch updates the lsu-lsi-drivers-plugin
VIB to resolve the following issue:
- PR 2242441: You might not be able to locate and turn on the LED for disks under lsi-msgpt35 controllers
You might not be able to locate and turn on the LED for disks under a lsi-msgpt35 controller.
This issue is resolved in this release. The fix introduces a new lsu-lsi-drivers-plugin which works with lsi-msgpt35 async and inbox drivers.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2270098 |
CVE numbers | N/A |
This patch updates the nvme
VIB to resolve the following issues:
- PR 2270098: ESXi hosts might fail with a purple diagnostic screen when using Intel NVMe devices
If a split I/O request is cancelled in the inbox nvme driver, the command completion path might unlock the queue lock due to an incorrect check of the command type. This might cause ESXi hosts to fail with a purple diagnostic screen.
This issue is resolved in this release.
- PR 2270098: You cannot see NVMe devices with doorbell stride different from 0h in ESXi hosts
VMware inbox nvme driver supports only doorbell stride with value 0h and you cannot use any other value or see NVMe devices with doorbell stride different from 0h.
This issue is resolved in this release.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | No |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2270098 |
CVE numbers | N/A |
This patch updates the vmware-esx-esxcli-nvme-plugin
VIB to resolve the following issue:
- PR 2270098: You might see incorrect NVMe device information by running the esxcfg-scsidevs –a command
For some Dell-branded NVMe devices, and for Marvell NR2241, you might see incorrect NVMe device information by running the
esxcfg-scsidevs –a
command or looking at the interface.This issue is resolved in this release.
Patch Category | Bugfix |
Patch Severity | Enhancement |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2272901 |
CVE numbers | N/A |
This patch updates the lsi-mr3
VIB.
- Upgrade to the Broadcom lsi-mr3 driver
The Broadcom lsi-mr3 driver is upgraded to version MR 7.8.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2275290 |
CVE numbers | N/A |
This patch updates the lsi-msgpt35
VIB.
- Update to the lsi-msgpt35 driver
The lsi-msgpt35 driver is updated to version 09.00.00.00-1vmw.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2275290 |
CVE numbers | N/A |
This patch updates the lsi-msgpt3
VIB.
- Update to the lsi-msgpt3 driver
The lsi-msgpt3 driver is updated to version 17.00.01.00-3vmw.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the net-vmxnet3
VIB.
Patch Category | Bugfix |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the native-misc-drivers
VIB.
Patch Category | Security |
Patch Severity | Important |
Host Reboot Required | Yes |
Virtual Machine Migration or Shutdown Required | Yes |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | 2228005, 2301046, 2232983, 2222019, 2224776, 2260077 |
CVE numbers | CVE-2019-5516, CVE-2019-5517, CVE-2019-5520 |
This patch updates the esx-base, vsan,
and vsanhealth
VIBs to resolve the following issues:
- PR 2224776: ESXi hosts might become unresponsive if the sfcbd service fails to process all forked processes
You might see many error messages such as
Heap globalCartel-1 already at its maximum size. Cannot expand
. in thevmkernel.log
, because the sfcbd service fails to process all forked processes. As a result, the ESXi host might become unresponsive and some operations might fail.This issue is resolved in this release.
- Update to the Network Time Protocol (NTP) daemon
The NTP daemon is updated to version ntp-4.2.8p12.
- Update to the libxml2 library
The ESXi userworld libxml2 library is updated to version 2.9.8.
- Update to the Python library
The Python third party library is updated to version 3.5.6.
- Update to the OpenSSH
The OpenSSH is updated to version 7.9 to resolve security issue with identifier CVE-2018-15473.
- Update to OpenSSL
The OpenSSL package is updated to version openssl-1.0.2r.
ESXi updates address an out-of-bounds vulnerability with the vertex shader functionality. Exploitation of this issue requires an attacker to have access to a virtual machine with 3D graphics enabled. Successful exploitation may lead to information disclosure or may allow attackers with normal user privileges to create a denial-of-service condition on their own VM. The workaround for this issue involves disabling the 3D-acceleration feature. This feature is not enabled by default on ESXi. The Common Vulnerabilities and Exposures project (cve.mitre.org) has assigned the identifier CVE-2019-5516 to this issue. See VMSA-2019-0006 for further information.
ESXi contains multiple out-of-bounds read vulnerabilities in the shader translator. Exploitation of these issues requires an attacker to have access to a virtual machine with 3D graphics enabled. Successful exploitation of these issues may lead to information disclosure or may allow attackers with normal user privileges to create a denial-of-service condition on their own VM. The workaround for these issues involves disabling the 3D-acceleration feature. This feature is not enabled by default on ESXi. The Common Vulnerabilities and Exposures project (cve.mitre.org) has assigned the identifier CVE-2019-5517 to these issues. See VMSA-2019-0006 for further information.
ESXi updates address an out-of-bounds read vulnerability. Exploitation of this issue requires an attacker to have access to a virtual machine with 3D graphics enabled. Successful exploitation of this issue may lead to information disclosure. The workaround for this issue involves disabling the 3D-acceleration feature. This feature is not enabled by default on ESXi. The Common Vulnerabilities and Exposures project (cve.mitre.org) has assigned the identifier CVE-2019-5520 to this issue. See VMSA-2019-0006 for further information.
Patch Category | Security |
Patch Severity | Important |
Host Reboot Required | No |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the tools-light
VIB.
Patch Category | Security |
Patch Severity | Important |
Host Reboot Required | No |
Virtual Machine Migration or Shutdown Required | No |
Affected Hardware | N/A |
Affected Software | N/A |
VIBs Included |
|
PRs Fixed | N/A |
CVE numbers | N/A |
This patch updates the esx-ui
VIB.
Profile Name | ESXi-6.7.0-20190402001-standard |
Build | For build information, see Patches Contained in this Release. |
Vendor | VMware, Inc. |
Release Date | April 11, 2019 |
Acceptance Level | PartnerSupported |
Affected Hardware | N/A |
Affected Software | N/A |
Affected VIBs |
|
PRs Fixed | 2227123, 2250472, 2273186, 2282052, 2267506, 2256771, 2269616, 2250697, 2193516, 2203431, 2264996, 2267158, 2231707, 2191342, 2197795, 2221935, 2193830, 2210364, 2177107, 2179263, 2185906, 2204508, 2211350, 2204204, 2226565, 2240385, 2163397, 2238134, 2213883, 2141221, 2240444, 2249713, 2184068, 2257480, 2235031, 2257354, 2212140, 2279897, 2227623, 2277648, 2221778, 2287560, 2268541, 2203837, 2236531, 2242656, 2250653, 2260748, 2256001, 2266762, 2263849, 2267698, 2271653, 2232538, 2210056, 2271322, 2272737, 2286563, 2178119, 2223879, 2221256, 2258272, 2219661, 2268826, 2224767, 2257590, 2241917, 2246891, 2228529, 2231435, 2226688, 2234599, 2256555, 2242441, 2270098, 2272901, 2275290 |
Related CVE numbers | N/A |
- This patch updates the following issues:
-
You might intermittently see alarms for very large values of counters for read or write disk latency, such as
datastore.totalReadLatency
anddatastore.totalWriteLatency
. -
When a Linux virtual machine is configured with specific memory size, such as 2052 MB or 2060 MB, instead of powering on, it might display a blank screen.
-
When you set up the network and enable vSphere Distributed Switch health check to perform configuration checks, the ESXi host might fail with purple diagnostic screen.
-
The SNMP agent might deliver an incorrect trap
vmwVmPoweredOn
when a virtual machine is selected under the Summary tab and the Snapshot Manager tab in the vSphere Web Client. -
The DVFilter might receive unexpected or corrupt values in the shared memory ring buffer, causing the internal function to return
NULL
. If thisNULL
value is not handled gracefully, an ESXi host might fail with purple diagnostic screen at DVFilter level. -
In the Events tab of the vSphere Client or the vSphere Web Client, when you power on or off virtual machines, you might see events similar to
Network connectivity restored on virtual switch XXX, portgroups: XXX. Physical NIC vmnicX is up
. When checking uplink status, the system report might include port groups that are not affected by the virtual machine powering on or off operation, which triggers the event. However, the event does not indicate a problem or prompt an action. -
If during the Windows Server Failover Cluster setup you change the default path policy from Fixed or Most Recently Used to Round Robin, the I/O of the cluster might fail and the cluster might stop responding.
-
In release builds, the smartd daemon might generate a lot of debugging and info messages into the syslog service logs.
-
The NIOC scheduler might reset the uplink network device if the uplink is rarely used. The reset is non-predictable.
-
ESXi hosts might intermittently fail if you disable global IPv6 addresses, because a code path still uses IPv6.
-
Due to a race condition, the virtual machine executable process might fail and shut down virtual machines during a reboot of the guest OS.
-
Hostd might become unresponsive while waiting for an external process initiated with the
esxcfg-syslog
command to finish. -
Large number of I/Os getting timeout on a heavy-loaded system might cause a soft lockup of physical CPUs. As a result, an ESXi host might fail with a purple diagnostic screen.
-
Invalid commands to a virtual SATA CD-ROM might trigger errors and increase the memory usage of virtual machines. This might lead to a failure of virtual machines if they are unable to allocate memory. You might see the following logs for invalid commands:
YYYY-MM-DDTHH:MM:SS.mmmZ| vcpu-1| I125: AHCI-USER: Invalid size in command
YYYY-MM-DDTHH:MM:SS.mmmZ| vcpu-1| I125: AHCI-VMM: sata0:0: Halted the port due to command abort.And a similar panic message:
YYYY-MM-DDTHH:MM:SS.mmmZ| vcpu-1| E105: PANIC: Unrecoverable memory allocation failure
-
If the Guest Introspection service in a vSphere 6.7 environment with more than 150 virtual machines is active, migration of virtual machines by using vSphere vMotion might fail with an error in the vSphere Web Client similar to
The source detected that the destination failed to resume
.
The destinationvmware.log
contains error messages similar to:2018-07-18T02:41:32.035Z| vmx| I125: MigrateSetState: Transitioning from state 11 to 12.
2018-07-18T02:41:32.035Z| vmx| I125: Migrate: Caching migration error message list:
2018-07-18T02:41:32.035Z| vmx| I125: [msg.checkpoint.migration.failedReceive] Failed to receive migration.
2018-07-18T02:41:32.035Z| vmx| I125: [msg.namespaceDb.badVersion] Incompatible version -1 (expect 2).
2018-07-18T02:41:32.035Z| vmx| I125: [msg.checkpoint.mrestoregroup.failed] An error occurred restoring the virtual machine state during migration.The vmkernel log contains error messages such as:
2018-07-18T02:32:43.011Z cpu5:66134)WARNING: Heap: 3534: Heap fcntlHeap-1 already at its maximum size. Cannot expand.
2018-07-18T02:41:35.613Z cpu2:66134)WARNING: Heap: 4169: Heap_Align(fcntlHeap-1, 200/200 bytes, 8 align) failed. caller: 0x41800aaca9a3 -
You can edit the statistical data to be collected only when the virtual machine is in a particular power state. For example, you cannot edit the memory data when the virtual machine is powered on. When you restart the hostd service, the virtual machine might not change its power state, which leads to false data initialization. The calculation of the host consumed memory counter might end up with division by zero and the chart might stop drawing graphs.
-
By default, the Isolation Response feature in an ESXi host with enabled vSphere HA is disabled. When the Isolation Response feature is enabled, a virtual machine port, which is connected to an NSX-T logical switch, might be blocked after a vSphere HA failover.
-
e1000 vNICs might intermittently drop unrecognized GRE packets due to a parsing issue.
-
While booting an Apple system, the boot process might stop with the following error message:
Shutting down firmware services...
Mutiboot buffer is too small.
Unrecoverable errorThe issue affects only certain Apple firmware versions.
-
If hostd restarts during a long-running quiesced snapshot operation, hostd might automatically run the snapshot Consolidation command to remove redundant disks and improve virtual machine performance. However, the Consolidation command might race with the running quiesced snapshot operation and cause failure of virtual machines.
-
The SNMP agent might display the status of all batteries as D_Failed when you run the
snmpwalk
command. This issue occurs due to the processing of code, which might not check the status properly and the misinterpretation of the compact sensors with sensor-specific code. -
When you use an IP hash or source MAC hash teaming policy, some packets from the packet list might use a different uplink. As a result, some of them might drop or not be sent out through the uplink determined by the teaming policy.
-
If an NFS datastore is configured as a syslog datastore and if the ESXi host disconnects from it, logging to the datastore stops and the ESXi host might become unresponsive.
-
When you replicate virtual machines by using vSphere Replication with Site Recovery Manager, the ESXi host might fail with a purple diagnostic screen immediately or within 24 hours. You might see a similar error:
PANIC bora/vmkernel/main/dlmalloc.c:4924 - Usage error in dlmalloc
. -
Larger configurations might exceed the limit for the number of locks and hostd starts failing with an error similar to:
hostd Panic: MXUserAllocSerialNumber: too many locks!
-
Тhe esxcli
$esxcli.network.nic.coalesce.set.Invoke
network command might return a false value as a result even though the value is correctly set. This might impact your automation scripts. -
When a virtual machine tries to access a VMFS datastore with corrupted directory, the ESXi host might fail with a purple diagnostic screen and the
DirEntry corruption
error message. -
The LACPPDU Actor key might change its value to 0 during the link status flapping.
-
Due to a timing issue in the VMkernel, buffers might not be flushed, and the heap gets exhausted. As a result, services such as hostd, vpxa and vmsyslogd might not be able to write logs on the ESXi host, and the host becomes unresponsive. In the
/var/log/vmkernel.log
, you might see a similar warning:WARNING: Heap: 3571: Heap vfat already at its maximum size. Cannot expand.
-
If you have modified the type of a virtual CD-ROM device of a virtual machine in a previous version of ESXi, after you update to a later version, when you try to install or update VMware Tools, the virtual machine might be terminated and marked as invalid.
-
When you renew a certificate, only the first certificate from the supplied chain of trust might be stored on the ESXi host. Any intermediate CA certificates are truncated. Because of the missing certificates, the chain to the root CA cannot be built. This leads to a warning for an untrusted connection.
-
If a vSphere API for Storage Awareness provider modifies the vSphere Virtual Volumes policy unattended, a null
VvolID
parameter might update the vSphere Virtual Volumes metadata. This results in a VASA call with a nullVvoId
parameter and a failure when creating a virtual machine snapshot. -
If a call from a vSphere API for Storage Awareness provider fails due to all connections to the virtual provider being busy, operations for parallel cloning of multiple virtual machines on a vSphere Virtual Volumes datastore might become unresponsive or fail with an error message similar to
Cannot complete file creation operation
. -
An ESXi host might fail with a purple diagnostic screen displaying a
sbflush_internal
panic message due to some discrepancies in the internal statistics. -
With ESXi 6.7 Update 2, you can migrate virtual machines by using vSphere vMotion between ESXi hosts with N-VDS and vSphere Standard Switches. To enable the feature, you must upgrade your vCenter Server system to vCenter Server 6.7 Update 2 and ESXi 6.7 Update 2 on both source and destination sites.
-
ESXi hosts might fail with a purple diagnostic screen due to a page fault exception. The error comes from a missing flag that could trigger cache eviction on an object that no longer exists, and result in a
NULL
pointer reference. -
When checking the uplink status, the teaming policy might not check if the uplink belongs to a port group and the affected port group might not be correct. As a result, you might see redundant VOB messages in the
/var/run/log/vobd.log
file, such as:Lost uplink redundancy on virtual switch "xxx". Physical NIC vmnicX is down. Affected portgroups:"xxx"
. -
You must manually add the claim rules to an ESXi host for Lenovo ThinkSystem DE Series Storage Arrays.
-
The vpxa service might fail due to excessive memory usage when iterating on a large directory tree. As a result, an ESXi host might get disconnected from a vCenter Server system.
-
The commands
esxcli software profile update
oresxcli software profile install
might fail on hosts running ESXi 6.7 Update 1 with enabled SecureBoot. You might see error messages such asFailed to setup upgrade using esx-update VIB
andFailed to mount tardisk in ramdisk: [Errno 1] Operation not permitted
. -
NFS volume mount might not persist after a reboot of an ESXi host due to intermittent failure in resolving the host name of the NFS server.
-
This issue happens if you create a VMK or vNIC port by using the Route Based on Physical NIC Load option when the physical NIC assigned to that port group is temporarily down. In such cases, the port team uplink is not set because there are no active uplinks in the portgroup. When the physical NIC becomes active again, teaming code fails to update the port data, resulting in a packet loss.
-
A race condition can occur when Plog Relog runs along with DecommissionMD, causing Relog to access freed Plog device state tables. This problem can cause a host failure with a purple diagnostic screen.
-
If you use devices with physical block size other than 4096 or 512 MB, in the vmkernel logs you might see many warnings similar to:
ScsiPath: 4395: The Physical block size "8192" reported by the path vmhba3:C0:T1:L0 is not supported.
-
After configuring or upgrading a vSAN stretched cluster, all cluster level requests might fail. You might see the following error message:
name '_VSAN_VER3' is not defined
. -
If the virtual machine folder is created using a name instead of UUID, the vSAN performance charts do not show the following VMDK metrics properly:
- IOPS and IOPS Limit
- Delayed Normalized IOPS -
In Host > Monitor > Performance > Physical Adapters > pNIC vSwitch Port Drop Rate you might see a counter showing dropped packets, while no packets are dropped. This is because the IOChain framework treats the reduced number of packets, when using LRO, as packet drops.
-
An ESXi host might fail with a purple diagnostic screen while powering on a virtual machine. This happens if the virtual machine name starts with “SRIOV” and has an uplink connected to a vSphere Distributed Switch.
-
In a rare race condition, when SNMP disables dynamic rulesets, but they are active in a Host Profile, you might see an unexpected compliance error such as
Ruleset dynamicruleset not found
. -
Recovery of a virtual machine running Windows 2008 or later by using vSphere Replication with application quiescing enabled might result in a corrupt replica. Corruption only happens when application quiescing is enabled. If VMware Tools on the virtual machine is not configured for application quiescing, vSphere Replication uses a lower level of consistency, such as file system quiescing, and replicas are not corrupted.
-
When you perform CIM requests more than 2 MB like firmware download, CIM service might fail on an ESXi host due to out of stack space and the
SFCB zdumb
file appears in/var/core
. -
The number of
syslog.log
archive files generated might be one less than the default value set by using thesyslog.global.defaultRotate
parameter. The number ofsyslog.log
archive files might also be different between ESXi versions 6.0 and 6.7. For instance, ifsyslog.global.defaultRotate
is set to 8 by default, ESXi 6.0 createssyslog.0.gz
tosyslog.7.gz
, while ESXi 6.7 createssyslog.0.gz
tosyslog.6.gz
. -
If you set the advanced config option
/Misc/MCEMonitorInterval
to a very large value, ESXi hosts might become unresponsive in around 10 days of uptime, due to a timer overflow. -
The P2MCache lock is a spin lock causing slow CPU performance that might cause starvation of other CPUs. As a result, an ESXi host might fail with a purple diagnostic screen and a
Spin count exceeded - possible deadlock, PVSCSIDoIOComplet, NFSAsyncIOComplete
error. -
An ESXi host might fail with a purple diagnostic screen during a Storage vMotion operation between NFS datastores, due to a race condition in the request completion path. The race condition happens under heavy I/O load in the NFS stack.
-
During disk decommission operation, a race condition in the decommission workflow might cause the ESXi host to fail with a purple diagnostic screen.
-
An ESXi host might fail with a purple diagnostic screen if RSS is enabled on a physical NIC. You might see a similar backtrace:
yyyy-mm-dd cpu9:2097316)0x451a8521bc70:[0x41801a68ad30]RSSPlugCleanupRSSEngine@(lb_netqueue_bal)#+0x7d stack: 0x4305e4c1a8e0, 0x41801a68af2b, 0x430db8c6b188, 0x4305e4c1a8e0,
0x0yyyy-mm-dd cpu9:2097316)0x451a8521bc90:[0x41801a68af2a]RSSPlugInitRSSEngine@(lb_netqueue_bal)#+0x127 stack: 0x0, 0x20c49ba5e353f7cf, 0x4305e4c1a930, 0x4305e4c1a800, 0x4305e4c1a9f0
yyyy-mm-dd cpu9:2097316)0x451a8521bcd0:[0x41801a68b21c]RSSPlug_PreBalanceWork@(lb_netqueue_bal)#+0x1cd stack: 0x32, 0x32, 0x0, 0xe0, 0x4305e4c1a800
yyyy-mm-dd cpu9:2097316)0x451a8521bd30:[0x41801a687752]Lb_PreBalanceWork@(lb_netqueue_bal)#+0x21f stack: 0x4305e4c1a800, 0xff, 0x0, 0x43057f38ea00, 0x4305e4c1a800
yyyy-mm-dd cpu9:2097316)0x451a8521bd80:[0x418019c13b18]UplinkNetqueueBal_BalanceCB@vmkernel#nover+0x6f1 stack: 0x4305e4bc6088, 0x4305e4c1a840, 0x4305e4c1a800, 0x4305e435fb30, 0x4305e4bc6088
yyyy-mm-dd cpu9:2097316)0x451a8521bf00:[0x418019cd1c89]UplinkAsyncProcessCallsHelperCB@vmkernel#nover+0x116 stack: 0x430749e60870, 0x430196737070, 0x451a85223000, 0x418019ae832b, 0x430749e60088
yyyy-mm-dd cpu9:2097316)0x451a8521bf30:[0x418019ae832a]HelperQueueFunc@vmkernel#nover+0x157 stack: 0x430749e600b8, 0x430749e600a8, 0x430749e600e0, 0x451a85223000, 0x430749e600b8
yyyy-mm-dd cpu9:2097316)0x451a8521bfe0:[0x418019d081f2]CpuSched_StartWorld@vmkernel#nover+0x77 stack: 0x0, 0x0, 0x0, 0x0, 0x0 -
When the VASA provider sends a rebind request to an ESXi host to switch the binding for a particular vSphere Virtual Volume driver, the ESXi host might switch the protocol endpoint and other resources to change the binding without any I/O disturbance. As a result, the ESXi host might fail with a purple diagnostic screen.
-
Some IPMI sensors might intermittently change from green to red and the reverse, which creates false hardware health alarms.
-
If you install an async 32-bit IMA plug-in and there are already installed 32-bit IMA plug-ins, the system might try to load these plug-ins and check if they are loaded from the QLogic independent iSCSI adapter. During this process the ESXi host might fail with a memory admission error.
-
If you run a reprotect operation by using the Site Recovery Manager, the operation might fail if a virtual machine is late with the initial synchronization. You might see an error similar to:
VR synchronization failed for VRM group {…}. Operation timed out: 10800 seconds
-
When I/O throttling is enabled on a protocol endpoint of Virtual Volumes on HPE 3PAR storage and if an I/O on a secondary LUN fails with an error
TASK_SET_FULL
, the I/O rate on all secondary LUNs that are associated with the protocol endpoint slows down. -
The qlnativefc driver is updated to version 3.1.8.0.
-
The nfnic driver is updated to version 4.0.0.17.
-
Emulex drivers might write logs at
/var/log/EMU/mili/mili2d.log
and fill up the 40 MB/var
file system logs of RAM drives. -
ESXi 6.7 Update 2 adds Solarflare native driver (sfvmk) support for Solarflare 10G and 40G network adaptor devices, such as SFN8542 and SFN8522.
-
The lsi-msgpt2 driver is updated to version 20.00.05.00 to address task management enhancements and critical bug fixes.
-
You might not be able to locate and turn on the LED for disks under a lsi-msgpt35 controller.
-
If a split I/O request is cancelled in the inbox nvme driver, the command completion path might unlock the queue lock due to an incorrect check of the command type. This might cause ESXi hosts to fail with a purple diagnostic screen.
-
VMware inbox nvme driver supports only doorbell stride with value 0h and you cannot use any other value or see NVMe devices with doorbell stride different from 0h.
-
For some Dell-branded NVMe devices, and for Marvell NR2241, you might see incorrect NVMe device information by running the
esxcfg-scsidevs –a
command or looking at the interface. -
The Broadcom lsi-mr3 driver is upgraded to version MR 7.8.
-
The lsi-msgpt35 driver is updated to version 09.00.00.00-1vmw.
-
The lsi-msgpt3 driver is updated to version 17.00.01.00-3vmw.
-
If you extract a Host Profile from an ESXi host with disabled IPv6 support, and the IPv4 default gateway is overridden while remediating that Host Profile, you might see a message that a virtual network adapter, such as DSwitch0, is to be created on the vSphere Distributed Switch (VDS), but the adapter is actually removed from the VDS. This is due to a "::" value of the
ipV6DefaultGateway
parameter. -
Some guest virtual machines might report their identity as an empty string if the guest is running a later version of VMware Tools than the ESXi host was released with, or if the host cannot recognize the
guestID
parameter of theGuestInfo
object. The issue affects mainly virtual machines with CentOS and VMware Photon OS. -
You might see repetitive messages in the vCenter Web Client, similar to
Delete host sub specification nsx
if you use network booting for ESXi hosts, due to some dummy code in hostd. -
The installation of the VMware Tools unmounts the ISO image file that ends with
autoinst.iso
, if it is already mounted to the virtual machine. To mount the VMware Tools ISO image file, you must click Install VMware Tools again. As a result, the virtual machine might stop responding. -
ESXi hosts might lose connectivity to the vCenter Sever system due to failure of the vpxa agent service in result of an invalid property update generated by the
PropertyCollector
object. A race condition in hostd leads to a malformed sequence of property update notification events that cause the invalid property update.
-
Profile Name | ESXi-6.7.0-20190402001-no-tools |
Build | For build information, see Patches Contained in this Release. |
Vendor | VMware, Inc. |
Release Date | April 11, 2019 |
Acceptance Level | PartnerSupported |
Affected Hardware | N/A |
Affected Software | N/A |
Affected VIBs |
|
PRs Fixed | 2227123, 2250472, 2273186, 2282052, 2267506, 2256771, 2269616, 2250697, 2193516, 2203431, 2264996, 2267158, 2231707, 2191342, 2197795, 2221935, 2193830, 2210364, 2177107, 2179263, 2185906, 2204508, 2211350, 2204204, 2226565, 2240385, 2163397, 2238134, 2213883, 2141221, 2240444, 2249713, 2184068, 2257480, 2235031, 2257354, 2212140, 2279897, 2227623, 2277648, 2221778, 2287560, 2268541, 2203837, 2236531, 2242656, 2250653, 2260748, 2256001, 2266762, 2263849, 2267698, 2271653, 2232538, 2210056, 2271322, 2272737, 2286563, 2178119, 2223879, 2221256, 2258272, 2219661, 2268826, 2224767, 2257590, 2241917, 2246891, 2228529, 2231435, 2226688, 2234599, 2256555, 2242441, 2270098, 2272901, 2275290 |
Related CVE numbers | N/A |
- This patch updates the following issues:
-
You might intermittently see alarms for very large values of counters for read or write disk latency, such as
datastore.totalReadLatency
anddatastore.totalWriteLatency
. -
When a Linux virtual machine is configured with specific memory size, such as 2052 MB or 2060 MB, instead of powering on, it might display a blank screen.
-
When you set up the network and enable vSphere Distributed Switch health check to perform configuration checks, the ESXi host might fail with purple diagnostic screen.
-
The SNMP agent might deliver an incorrect trap
vmwVmPoweredOn
when a virtual machine is selected under the Summary tab and the Snapshot Manager tab in the vSphere Web Client. -
The DVFilter might receive unexpected or corrupt values in the shared memory ring buffer, causing the internal function to return
NULL
. If thisNULL
value is not handled gracefully, an ESXi host might fail with purple diagnostic screen at DVFilter level. -
In the Events tab of the vSphere Client or the vSphere Web Client, when you power on or off virtual machines, you might see events similar to
Network connectivity restored on virtual switch XXX, portgroups: XXX. Physical NIC vmnicX is up
. When checking uplink status, the system report might include port groups that are not affected by the virtual machine powering on or off operation, which triggers the event. However, the event does not indicate a problem or prompt an action. -
If during the Windows Server Failover Cluster setup you change the default path policy from Fixed or Most Recently Used to Round Robin, the I/O of the cluster might fail and the cluster might stop responding.
-
In release builds, the smartd daemon might generate a lot of debugging and info messages into the syslog service logs.
-
The NIOC scheduler might reset the uplink network device if the uplink is rarely used. The reset is non-predictable.
-
ESXi hosts might intermittently fail if you disable global IPv6 addresses, because a code path still uses IPv6.
-
Due to a race condition, the virtual machine executable process might fail and shut down virtual machines during a reboot of the guest OS.
-
Hostd might become unresponsive while waiting for an external process initiated with the
esxcfg-syslog
command to finish. -
Large number of I/Os getting timeout on a heavy-loaded system might cause a soft lockup of physical CPUs. As a result, an ESXi host might fail with a purple diagnostic screen.
-
Invalid commands to a virtual SATA CD-ROM might trigger errors and increase the memory usage of virtual machines. This might lead to a failure of virtual machines if they are unable to allocate memory. You might see the following logs for invalid commands:
YYYY-MM-DDTHH:MM:SS.mmmZ| vcpu-1| I125: AHCI-USER: Invalid size in command
YYYY-MM-DDTHH:MM:SS.mmmZ| vcpu-1| I125: AHCI-VMM: sata0:0: Halted the port due to command abort.And a similar panic message:
YYYY-MM-DDTHH:MM:SS.mmmZ| vcpu-1| E105: PANIC: Unrecoverable memory allocation failure
-
If the Guest Introspection service in a vSphere 6.7 environment with more than 150 virtual machines is active, migration of virtual machines by using vSphere vMotion might fail with an error in the vSphere Web Client similar to
The source detected that the destination failed to resume
.
The destinationvmware.log
contains error messages similar to:2018-07-18T02:41:32.035Z| vmx| I125: MigrateSetState: Transitioning from state 11 to 12.
2018-07-18T02:41:32.035Z| vmx| I125: Migrate: Caching migration error message list:
2018-07-18T02:41:32.035Z| vmx| I125: [msg.checkpoint.migration.failedReceive] Failed to receive migration.
2018-07-18T02:41:32.035Z| vmx| I125: [msg.namespaceDb.badVersion] Incompatible version -1 (expect 2).
2018-07-18T02:41:32.035Z| vmx| I125: [msg.checkpoint.mrestoregroup.failed] An error occurred restoring the virtual machine state during migration.The vmkernel log contains error messages such as:
2018-07-18T02:32:43.011Z cpu5:66134)WARNING: Heap: 3534: Heap fcntlHeap-1 already at its maximum size. Cannot expand.
2018-07-18T02:41:35.613Z cpu2:66134)WARNING: Heap: 4169: Heap_Align(fcntlHeap-1, 200/200 bytes, 8 align) failed. caller: 0x41800aaca9a3 -
You can edit the statistical data to be collected only when the virtual machine is in a particular power state. For example, you cannot edit the memory data when the virtual machine is powered on. When you restart the hostd service, the virtual machine might not change its power state, which leads to false data initialization. The calculation of the host consumed memory counter might end up with division by zero and the chart might stop drawing graphs.
-
By default, the Isolation Response feature in an ESXi host with enabled vSphere HA is disabled. When the Isolation Response feature is enabled, a virtual machine port, which is connected to an NSX-T logical switch, might be blocked after a vSphere HA failover.
-
e1000 vNICs might intermittently drop unrecognized GRE packets due to a parsing issue.
-
While booting an Apple system, the boot process might stop with the following error message:
Shutting down firmware services...
Mutiboot buffer is too small.
Unrecoverable errorThe issue affects only certain Apple firmware versions.
-
If hostd restarts during a long-running quiesced snapshot operation, hostd might automatically run the snapshot Consolidation command to remove redundant disks and improve virtual machine performance. However, the Consolidation command might race with the running quiesced snapshot operation and cause failure of virtual machines.
-
The SNMP agent might display the status of all batteries as D_Failed when you run the
snmpwalk
command. This issue occurs due to the processing of code, which might not check the status properly and the misinterpretation of the compact sensors with sensor-specific code. -
When you use an IP hash or source MAC hash teaming policy, some packets from the packet list might use a different uplink. As a result, some of them might drop or not be sent out through the uplink determined by the teaming policy.
-
If an NFS datastore is configured as a syslog datastore and if the ESXi host disconnects from it, logging to the datastore stops and the ESXi host might become unresponsive.
-
When you replicate virtual machines by using vSphere Replication with Site Recovery Manager, the ESXi host might fail with a purple diagnostic screen immediately or within 24 hours. You might see a similar error:
PANIC bora/vmkernel/main/dlmalloc.c:4924 - Usage error in dlmalloc
. -
Larger configurations might exceed the limit for the number of locks and hostd starts failing with an error similar to:
hostd Panic: MXUserAllocSerialNumber: too many locks!
-
Тhe esxcli
$esxcli.network.nic.coalesce.set.Invoke
network command might return a false value as a result even though the value is correctly set. This might impact your automation scripts. -
When a virtual machine tries to access a VMFS datastore with corrupted directory, the ESXi host might fail with a purple diagnostic screen and the
DirEntry corruption
error message. -
The LACPPDU Actor key might change its value to 0 during the link status flapping.
-
Due to a timing issue in the VMkernel, buffers might not be flushed, and the heap gets exhausted. As a result, services such as hostd, vpxa and vmsyslogd might not be able to write logs on the ESXi host, and the host becomes unresponsive. In the
/var/log/vmkernel.log
, you might see a similar warning:WARNING: Heap: 3571: Heap vfat already at its maximum size. Cannot expand.
-
If you have modified the type of a virtual CD-ROM device of a virtual machine in a previous version of ESXi, after you update to a later version, when you try to install or update VMware Tools, the virtual machine might be terminated and marked as invalid.
-
When you renew a certificate, only the first certificate from the supplied chain of trust might be stored on the ESXi host. Any intermediate CA certificates are truncated. Because of the missing certificates, the chain to the root CA cannot be built. This leads to a warning for an untrusted connection.
-
If a vSphere API for Storage Awareness provider modifies the vSphere Virtual Volumes policy unattended, a null
VvolID
parameter might update the vSphere Virtual Volumes metadata. This results in a VASA call with a nullVvoId
parameter and a failure when creating a virtual machine snapshot. -
If a call from a vSphere API for Storage Awareness provider fails due to all connections to the virtual provider being busy, operations for parallel cloning of multiple virtual machines on a vSphere Virtual Volumes datastore might become unresponsive or fail with an error message similar to
Cannot complete file creation operation
. -
An ESXi host might fail with a purple diagnostic screen displaying a
sbflush_internal
panic message due to some discrepancies in the internal statistics. -
With ESXi 6.7 Update 2, you can migrate virtual machines by using vSphere vMotion between ESXi hosts with N-VDS and vSphere Standard Switches. To enable the feature, you must upgrade your vCenter Server system to vCenter Server 6.7 Update 2 and ESXi 6.7 Update 2 on both source and destination sites.
-
ESXi hosts might fail with a purple diagnostic screen due to a page fault exception. The error comes from a missing flag that could trigger cache eviction on an object that no longer exists, and result in a
NULL
pointer reference. -
When checking the uplink status, the teaming policy might not check if the uplink belongs to a port group and the affected port group might not be correct. As a result, you might see redundant VOB messages in the
/var/run/log/vobd.log
file, such as:Lost uplink redundancy on virtual switch "xxx". Physical NIC vmnicX is down. Affected portgroups:"xxx"
. -
You must manually add the claim rules to an ESXi host for Lenovo ThinkSystem DE Series Storage Arrays.
-
The vpxa service might fail due to excessive memory usage when iterating on a large directory tree. As a result, an ESXi host might get disconnected from a vCenter Server system.
-
The commands
esxcli software profile update
oresxcli software profile install
might fail on hosts running ESXi 6.7 Update 1 with enabled SecureBoot. You might see error messages such asFailed to setup upgrade using esx-update VIB
andFailed to mount tardisk in ramdisk: [Errno 1] Operation not permitted
. -
NFS volume mount might not persist after a reboot of an ESXi host due to intermittent failure in resolving the host name of the NFS server.
-
This issue happens if you create a VMK or vNIC port by using the Route Based on Physical NIC Load option when the physical NIC assigned to that port group is temporarily down. In such cases, the port team uplink is not set because there are no active uplinks in the portgroup. When the physical NIC becomes active again, teaming code fails to update the port data, resulting in a packet loss.
-
A race condition can occur when Plog Relog runs along with DecommissionMD, causing Relog to access freed Plog device state tables. This problem can cause a host failure with a purple diagnostic screen.
-
If you use devices with physical block size other than 4096 or 512 MB, in the vmkernel logs you might see many warnings similar to:
ScsiPath: 4395: The Physical block size "8192" reported by the path vmhba3:C0:T1:L0 is not supported.
-
After configuring or upgrading a vSAN stretched cluster, all cluster level requests might fail. You might see the following error message:
name '_VSAN_VER3' is not defined
. -
If the virtual machine folder is created using a name instead of UUID, the vSAN performance charts do not show the following VMDK metrics properly:
- IOPS and IOPS Limit
- Delayed Normalized IOPS -
In Host > Monitor > Performance > Physical Adapters > pNIC vSwitch Port Drop Rate you might see a counter showing dropped packets, while no packets are dropped. This is because the IOChain framework treats the reduced number of packets, when using LRO, as packet drops.
-
An ESXi host might fail with a purple diagnostic screen while powering on a virtual machine. This happens if the virtual machine name starts with “SRIOV” and has an uplink connected to a vSphere Distributed Switch.
-
In a rare race condition, when SNMP disables dynamic rulesets, but they are active in a Host Profile, you might see an unexpected compliance error such as
Ruleset dynamicruleset not found
. -
Recovery of a virtual machine running Windows 2008 or later by using vSphere Replication with application quiescing enabled might result in a corrupt replica. Corruption only happens when application quiescing is enabled. If VMware Tools on the virtual machine is not configured for application quiescing, vSphere Replication uses a lower level of consistency, such as file system quiescing, and replicas are not corrupted.
-
When you perform CIM requests more than 2 MB like firmware download, CIM service might fail on an ESXi host due to out of stack space and the
SFCB zdumb
file appears in/var/core
. -
The number of
syslog.log
archive files generated might be one less than the default value set by using thesyslog.global.defaultRotate
parameter. The number ofsyslog.log
archive files might also be different between ESXi versions 6.0 and 6.7. For instance, ifsyslog.global.defaultRotate
is set to 8 by default, ESXi 6.0 createssyslog.0.gz
tosyslog.7.gz
, while ESXi 6.7 createssyslog.0.gz
tosyslog.6.gz
. -
If you set the advanced config option
/Misc/MCEMonitorInterval
to a very large value, ESXi hosts might become unresponsive in around 10 days of uptime, due to a timer overflow. -
The P2MCache lock is a spin lock causing slow CPU performance that might cause starvation of other CPUs. As a result, an ESXi host might fail with a purple diagnostic screen and a
Spin count exceeded - possible deadlock, PVSCSIDoIOComplet, NFSAsyncIOComplete
error. -
An ESXi host might fail with a purple diagnostic screen during a Storage vMotion operation between NFS datastores, due to a race condition in the request completion path. The race condition happens under heavy I/O load in the NFS stack.
-
During disk decommission operation, a race condition in the decommission workflow might cause the ESXi host to fail with a purple diagnostic screen.
-
An ESXi host might fail with a purple diagnostic screen if RSS is enabled on a physical NIC. You might see a similar backtrace:
yyyy-mm-dd cpu9:2097316)0x451a8521bc70:[0x41801a68ad30]RSSPlugCleanupRSSEngine@(lb_netqueue_bal)#+0x7d stack: 0x4305e4c1a8e0, 0x41801a68af2b, 0x430db8c6b188, 0x4305e4c1a8e0,
0x0yyyy-mm-dd cpu9:2097316)0x451a8521bc90:[0x41801a68af2a]RSSPlugInitRSSEngine@(lb_netqueue_bal)#+0x127 stack: 0x0, 0x20c49ba5e353f7cf, 0x4305e4c1a930, 0x4305e4c1a800, 0x4305e4c1a9f0
yyyy-mm-dd cpu9:2097316)0x451a8521bcd0:[0x41801a68b21c]RSSPlug_PreBalanceWork@(lb_netqueue_bal)#+0x1cd stack: 0x32, 0x32, 0x0, 0xe0, 0x4305e4c1a800
yyyy-mm-dd cpu9:2097316)0x451a8521bd30:[0x41801a687752]Lb_PreBalanceWork@(lb_netqueue_bal)#+0x21f stack: 0x4305e4c1a800, 0xff, 0x0, 0x43057f38ea00, 0x4305e4c1a800
yyyy-mm-dd cpu9:2097316)0x451a8521bd80:[0x418019c13b18]UplinkNetqueueBal_BalanceCB@vmkernel#nover+0x6f1 stack: 0x4305e4bc6088, 0x4305e4c1a840, 0x4305e4c1a800, 0x4305e435fb30, 0x4305e4bc6088
yyyy-mm-dd cpu9:2097316)0x451a8521bf00:[0x418019cd1c89]UplinkAsyncProcessCallsHelperCB@vmkernel#nover+0x116 stack: 0x430749e60870, 0x430196737070, 0x451a85223000, 0x418019ae832b, 0x430749e60088
yyyy-mm-dd cpu9:2097316)0x451a8521bf30:[0x418019ae832a]HelperQueueFunc@vmkernel#nover+0x157 stack: 0x430749e600b8, 0x430749e600a8, 0x430749e600e0, 0x451a85223000, 0x430749e600b8
yyyy-mm-dd cpu9:2097316)0x451a8521bfe0:[0x418019d081f2]CpuSched_StartWorld@vmkernel#nover+0x77 stack: 0x0, 0x0, 0x0, 0x0, 0x0 -
When the VASA provider sends a rebind request to an ESXi host to switch the binding for a particular vSphere Virtual Volume driver, the ESXi host might switch the protocol endpoint and other resources to change the binding without any I/O disturbance. As a result, the ESXi host might fail with a purple diagnostic screen.
-
Some IPMI sensors might intermittently change from green to red and the reverse, which creates false hardware health alarms.
-
If you install an async 32-bit IMA plug-in and there are already installed 32-bit IMA plug-ins, the system might try to load these plug-ins and check if they are loaded from the QLogic independent iSCSI adapter. During this process the ESXi host might fail with a memory admission error.
-
If you run a reprotect operation by using the Site Recovery Manager, the operation might fail if a virtual machine is late with the initial synchronization. You might see an error similar to:
VR synchronization failed for VRM group {…}. Operation timed out: 10800 seconds
-
When I/O throttling is enabled on a protocol endpoint of Virtual Volumes on HPE 3PAR storage and if an I/O on a secondary LUN fails with an error
TASK_SET_FULL
, the I/O rate on all secondary LUNs that are associated with the protocol endpoint slows down. -
The qlnativefc driver is updated to version 3.1.8.0.
-
The nfnic driver is updated to version 4.0.0.17.
-
Emulex drivers might write logs at
/var/log/EMU/mili/mili2d.log
and fill up the 40 MB/var
file system logs of RAM drives. -
ESXi 6.7 Update 2 adds Solarflare native driver (sfvmk) support for Solarflare 10G and 40G network adaptor devices, such as SFN8542 and SFN8522.
-
The lsi-msgpt2 driver is updated to version 20.00.05.00 to address task management enhancements and critical bug fixes.
-
You might not be able to locate and turn on the LED for disks under a lsi-msgpt35 controller.
-
If a split I/O request is cancelled in the inbox nvme driver, the command completion path might unlock the queue lock due to an incorrect check of the command type. This might cause ESXi hosts to fail with a purple diagnostic screen.
-
VMware inbox nvme driver supports only doorbell stride with value 0h and you cannot use any other value or see NVMe devices with doorbell stride different from 0h.
-
For some Dell-branded NVMe devices, and for Marvell NR2241, you might see incorrect NVMe device information by running the
esxcfg-scsidevs –a
command or looking at the interface. -
The Broadcom lsi-mr3 driver is upgraded to version MR 7.8.
-
The lsi-msgpt35 driver is updated to version 09.00.00.00-1vmw.
-
The lsi-msgpt3 driver is updated to version 17.00.01.00-3vmw.
-
If you extract a Host Profile from an ESXi host with disabled IPv6 support, and the IPv4 default gateway is overridden while remediating that Host Profile, you might see a message that a virtual network adapter, such as DSwitch0, is to be created on the vSphere Distributed Switch (VDS), but the adapter is actually removed from the VDS. This is due to a "::" value of the
ipV6DefaultGateway
parameter. -
Some guest virtual machines might report their identity as an empty string if the guest is running a later version of VMware Tools than the ESXi host was released with, or if the host cannot recognize the
guestID
parameter of theGuestInfo
object. The issue affects mainly virtual machines with CentOS and VMware Photon OS. -
You might see repetitive messages in the vCenter Web Client, similar to
Delete host sub specification nsx
if you use network booting for ESXi hosts, due to some dummy code in hostd. -
The installation of the VMware Tools unmounts the ISO image file that ends with
autoinst.iso
, if it is already mounted to the virtual machine. To mount the VMware Tools ISO image file, you must click Install VMware Tools again. As a result, the virtual machine might stop responding. -
ESXi hosts might lose connectivity to the vCenter Sever system due to failure of the vpxa agent service in result of an invalid property update generated by the
PropertyCollector
object. A race condition in hostd leads to a malformed sequence of property update notification events that cause the invalid property update.
-
Profile Name | ESXi-6.7.0-20190401001s-standard |
Build | For build information, see Patches Contained in this Release. |
Vendor | VMware, Inc. |
Release Date | Аpril 11, 2019 |
Acceptance Level | PartnerSupported |
Affected Hardware | N/A |
Affected Software | N/A |
Affected VIBs |
|
PRs Fixed | 2228005, 2301046, 2232983, 2222019, 2224776, 2260077, 2232955 |
Related CVE numbers | CVE-2019-5516, CVE-2019-5517, CVE-2019-5520 |
This patch updates the following issues:
-
-
You might see many error messages such as
Heap globalCartel-1 already at its maximum size. Cannot expand
. in thevmkernel.log
, because the sfcbd service fails to process all forked processes. As a result, the ESXi host might become unresponsive and some operations might fail. -
The NTP daemon is updated to version ntp-4.2.8p12.
-
The ESXi userworld libxml2 library is updated to version 2.9.8.
-
The Python third party library is updated to version 3.5.6.
-
The OpenSSH is updated to version 7.9 to resolve security issue with identifier CVE-2018-15473.
-
The OpenSSL package is updated to version openssl-1.0.2r.
-
ESXi updates address an out-of-bounds vulnerability with the vertex shader functionality. Exploitation of this issue requires an attacker to have access to a virtual machine with 3D graphics enabled. Successful exploitation may lead to information disclosure or may allow attackers with normal user privileges to create a denial-of-service condition on their own VM. The workaround for this issue involves disabling the 3D-acceleration feature. This feature is not enabled by default on ESXi. The Common Vulnerabilities and Exposures project (cve.mitre.org) has assigned the identifier CVE-2019-5516 to this issue. See VMSA-2019-0006 for further information.
-
ESXi contains multiple out-of-bounds read vulnerabilities in the shader translator. Exploitation of these issues requires an attacker to have access to a virtual machine with 3D graphics enabled. Successful exploitation of these issues may lead to information disclosure or may allow attackers with normal user privileges to create a denial-of-service condition on their own VM. The workaround for these issues involves disabling the 3D-acceleration feature. This feature is not enabled by default on ESXi. The Common Vulnerabilities and Exposures project (cve.mitre.org) has assigned the identifier CVE-2019-5517 to these issues. See VMSA-2019-0006 for further information.
-
ESXi updates address an out-of-bounds read vulnerability. Exploitation of this issue requires an attacker to have access to a virtual machine with 3D graphics enabled. Successful exploitation of this issue may lead to information disclosure. The workaround for this issue involves disabling the 3D-acceleration feature. This feature is not enabled by default on ESXi. The Common Vulnerabilities and Exposures project (cve.mitre.org) has assigned the identifier CVE-2019-5520 to this issue. See VMSA-2019-0006 for further information.
-
Profile Name | ESXi-6.7.0-20190301001s-no-tools |
Build | For build information, see Patches Contained in this Release. |
Vendor | VMware, Inc. |
Release Date | April 11, 2019 |
Acceptance Level | PartnerSupported |
Affected Hardware | N/A |
Affected Software | N/A |
Affected VIBs |
|
PRs Fixed | 2228005, 2301046, 2232983, 2222019, 2224776, 2260077, 2232955 |
Related CVE numbers | CVE-2019-5516, CVE-2019-5517, CVE-2019-5520 |
This patch updates the following issues:
-
-
You might see many error messages such as
Heap globalCartel-1 already at its maximum size. Cannot expand
. in thevmkernel.log
, because the sfcbd service fails to process all forked processes. As a result, the ESXi host might become unresponsive and some operations might fail. -
The NTP daemon is updated to version ntp-4.2.8p12.
-
The ESXi userworld libxml2 library is updated to version 2.9.8.
-
The Python third party library is updated to version 3.5.6.
-
The OpenSSH is updated to version 7.9 to resolve security issue with identifier CVE-2018-15473.
-
The OpenSSL package is updated to version openssl-1.0.2r.
-
ESXi updates address an out-of-bounds vulnerability with the vertex shader functionality. Exploitation of this issue requires an attacker to have access to a virtual machine with 3D graphics enabled. Successful exploitation may lead to information disclosure or may allow attackers with normal user privileges to create a denial-of-service condition on their own VM. The workaround for this issue involves disabling the 3D-acceleration feature. This feature is not enabled by default on ESXi. The Common Vulnerabilities and Exposures project (cve.mitre.org) has assigned the identifier CVE-2019-5516 to this issue. See VMSA-2019-0006 for further information.
-
ESXi contains multiple out-of-bounds read vulnerabilities in the shader translator. Exploitation of these issues requires an attacker to have access to a virtual machine with 3D graphics enabled. Successful exploitation of these issues may lead to information disclosure or may allow attackers with normal user privileges to create a denial-of-service condition on their own VM. The workaround for these issues involves disabling the 3D-acceleration feature. This feature is not enabled by default on ESXi. The Common Vulnerabilities and Exposures project (cve.mitre.org) has assigned the identifier CVE-2019-5517 to these issues. See VMSA-2019-0006 for further information.
-
ESXi updates address an out-of-bounds read vulnerability. Exploitation of this issue requires an attacker to have access to a virtual machine with 3D graphics enabled. Successful exploitation of this issue may lead to information disclosure. The workaround for this issue involves disabling the 3D-acceleration feature. This feature is not enabled by default on ESXi. The Common Vulnerabilities and Exposures project (cve.mitre.org) has assigned the identifier CVE-2019-5520 to this issue. See VMSA-2019-0006 for further information.
-
Known Issues
The known issues are grouped as follows.
Miscellaneous- The ESXi Syslog service might display inaccurate timestamps at start or rotation
The ESXi Syslog service might display inaccurate timestamps when the service starts or restarts, or when the log reaches its maximum size by configuration and rotates.
Workaround: This issue is resolved for fresh installs of ESXi. To fix the issue on the existing configuration for the
hostd.log
, you must:- Open the
/etc/vmsyslog.conf.d/hostd.conf
file. - Replace
onrotate = logger -t Hostd < /var/run/vmware/hostdLogHeader.txt
with
onrotate = printf '%%s - last log rotation time, %%s\n' "$(date --utc +%%FT%%T.%%3NZ)" "$(cat /var/run/vmware/hostdLogHeader.txt)" | logger -t Hostd
- Save the changes.
- Restart the vmsyslogd service by running
/etc/init.d/vmsyslogd restart.
- Open the
- You cannot migrate virtual machines by using vSphere vMotion on different NSX Logical Switches
In ESXi 6.7 Update 2, migration of virtual machines using vSphere vMotion between two different NSX Logical Switches is not supported. You can still migrate virtual machines with vSphere vMotion on the same NSX Logical Switch.
Workaround: None
- When you upgrade to VMware ESXi 6.7 Update 2, the SSH service might stop responding
When you upgrade to VMware ESXi 6.7 Update 2 and boot an ESXi host, the
/etc/ssh/sshd_config
file might contain Unicode characters. As a result, the SSH configuration might not be updated and the SSH service might stop responding, refusing the incoming connections.Workaround: You must disable and then enable the SSH service manually, following the steps in VMware knowledge base article 2004746.
When you upgrade to VMware ESXi 6.7 Update 2, the SSH update is successful if the first line in the/etc/ssh/sshd_config
file displays:# Version 6.7.2.0
. - Linux guest OS in ESXi hosts might become unresponsive if you run the kdump utility while vIOMMU is enabled
Some Linux guest OS might become unresponsive while initializing the kdump kernel if vIOMMU is enabled.
Workaround: If kdump is required, turn off interrupt remapping in the guest OS by appending
intremap=off
to the grub menu or disable vIOMMU for this virtual machine completely.