ESXi - /var/log/vmkernel.log
yyyy-mm-dd cpu5:2924874)MemSched: 14642: uw.2924860 (14074759) extraMin/extraFromParent: 256/256, likewise (790) childEmin/eMinLimit: 99871/100096yyyy-mm-dd cpu5:2924874)MemSched: 14635: Admission failure in path: likewise/lwsmd.2924860/uw.2924860yyyy-mm-dd cpu5:2924874)MemSched: 14642: uw.2924860 (14074759) extraMin/extraFromParent: 256/256, likewise (790) childEmin/eMinLimit: 99871/100096yyyy-mm-dd cpu5:2924874)MemSched: 14635: Admission failure in path: likewise/lwsmd.2924860/uw.2924860
yyyy-mm-dd cpu11:2100089)ALERT: hostd detected to be non-responsive
.259Z Wa(180) vmkwarning: cpu27:2101383)WARNING: LinuxThread: 411: lwsmd: Error cloning thread: -12 (bad0014)YYYY-MM-DDTHH:MM:SS.259Z In(182) vmkernel: cpu27:2101383)Admission failure in path: host/vim/vmvisor/likewise:lwsmd.2101378:uwWorldStore.2101378YYYY-MM-DDTHH:MM:SS.259Z In(182) vmkernel: cpu27:2101383)uwWorldStore.2101378 (10580) requires 44 KB, asked 44 KB from likewise (836) which has 93184 KB occupied and 0 KB available.YYYY-MM-DDTHH:MM:SS.259Z In(182) vmkernel: cpu27:2101383)Admission failure in path: host/vim/vmvisor/likewise:lwsmd.2101378:uwWorldStore.2101378YYYY-MM-DDTHH:MM:SS.259Z In(182) vmkernel: cpu27:2101383)uwWorldStore.2101378 (10580) requires 44 KB, asked 44 KB from likewise (836) which has 93184 KB occupied and 0 KB available.YYYY-MM-DDTHH:MM:SS.259Z Wa(180) vmkwarning: cpu27:2101383)WARNING: World: 3234: Could not allocate new world handle for world ID: 26953837: Out of memoryYYYY-MM-DDTHH:MM:SS
.938Z Wa(180) vmkwarning: cpu52:2135899)WARNING: MemSchedAdmit: 1263: Group likewise: Requested memory limit 0 KB insufficient to support effective reservation 25192 KB
YYYY-MM-DDTHH:MM:SS.395Z In(182) vmkernel: cpu6:2135943)User: 3259: lwsmd: wantCoreDump:lwsmd signal:11 exitCode:0 coredump:enabledYYYY-MM-DDTHH:MM:SS.526Z In(182) vmkernel: cpu6:2135943)UserDump: 3157: lwsmd: Dumping cartel 2135924 (from world 2135943) to file /var/core/lwsmd-zdump.000 ...YYYY-MM-DDTHH:MM:SS
/var/log/vobd.log fileYYYY-MM-DDTHH:MM:SS.810Z In(14) vobd[2097766]: [UserWorldCorrelator] 10473605981us: [esx.problem.application.core.dumped] An application (/usr/lib/vmware/likewise/sbin/lwsmd) running on ESXi host has crashed (1 time(s) so far). A core file may have been created at /var/core/lwsmd-zdump.000
YYYY-MM-DDTHH:MM:SS.216Z Er(27) lwsmd[2100830]: [lwsm] Could not start bootstrap service: LW_ERROR_SERVICE_UNRESPONSIVE
VMware ESXi 7.x
VMware ESXi 8.x
The issue occurs due to the exhaustion of Likewise memory because of existing memory leaks in Active Directory operations and related libraries. When smart card authentication is enabled and configured on the ESXi hosts, Likewise memory leaks are also observed.
Option 1: Non-persistent workaround
The below script monitors the Likewise memory usage and generates a lwis-oom-stats file in the /tmp directory with Likewise memory stats. When it detects that Likewise runs out of memory, it will automatically remediate the issue.
Note: This workaround is only temporary and will not persist if the ESXi host is rebooted.
chmod +x /tmp/lwis-mem-check-2.shsetsid /tmp/lwis-mem-check-2.sh >/dev/null 2>&1 < /dev/null &
Option 2: Persistent workaround
Important: Ensure to back up the ESXi host configuration before proceeding. Refer to KB How to back up and restore the ESXi host configuration.
Note: This is not an option if you have secureboot enabled.
If unfamiliar with the below steps, contact VMware support to assist.
Note: You will need a shared datastore path where you can place the lwis-mem-check.sh script that can be retrieved during startup.
cp /etc/rc.local.d/local.sh /local.sh
vi /etc/rc.local.d/local.sh
cp /vmfs/volumes/<Datastore>/<script>/lwis-mem-check-2.sh /tmp/<Datastore>/<script> should be replaced with the correct datastore directory path where the script is located.auto-backup.shNotes:
Option 3
If the above option is not valid, remove the host from active directory and stop the Likewise service on the ESXi host: /etc/init.d/lwsmd stop
If the lwsmd service is enabled to start automatically, update the service policy to "Start and stop manually."
This issue is being checked by Diagnostics for VMware Cloud Foundation.
The check is as follows: