Hello,
I recently restarted the management agents (using the VMWare tutorial) in order to solve some issues. However, that did not correct the issues, instead it has introduced more problems : Now, I cannot connect to the ESXi server with vSphere Client 5.0. I get the following error :
vSphere Client could not connect to "myesxiiserver".
An unknown connection error occurred. (The request failed because of a connection failure. (Unable to connect to the remote server)).
Moreover, when I try to access to https://myesxiipserver/, I get a 404 HTTP error.
But, I can access to the ESXi server by SSH and I can telnet the server with the port 902.
Does anybody know how I can fix this issue ?
Thanks.
Cheers and happy new year
Hi ,
When you take DCUI console to the host what are we getting in alt+f12 screen ? Prior to that i also request you to check the management network from the same console ie do a test and let me know the result
Note:How did you confirm management agent restart was Successful or not ?
Regard's
Sree
EMCIE|EMCISA-V2|EMCISA|VCP4|VCP5|MCP|MCTS AD2008
This is what I get when I ALT+F12 from the DCUI console :
Prior to that i also request you to check the management network from the same console ie do a test and let me know the result
From the same console, the test to check managemement network is successful :
Note:How did you confirm management agent restart was Successful or not ?
The DCUI console displayed an "OK" in the screen. I assume the restart was successful.
I tried to restart directly by SSH. Here are the result :
~ # services.sh restart
Running sfcbd stop
This operation is not supported.
Please use /etc/init.d/sfcbd-watchdog stop
Running wsman stop
Stopping openwsmand
Running sfcbd-watchdog stop
Running vpxa stop
watchdog-vpxa: Terminating watchdog process with PID 3262937
vpxa stopped.
Running vobd stop
watchdog-vobd: Terminating watchdog process with PID 3264951
vobd stopped
Running cdp stop
watchdog-cdp: Terminating watchdog process with PID 3266980
Running dcbd stop
watchdog-dcbd: Terminating watchdog process with PID 3264913
Running memscrubd stop
memscrubd is not running
Running slpd stop
Stopping slpd
Running sensord stop
sensord is not running
Running storageRM stop
watchdog-storageRM: Terminating watchdog process with PID 3266857
storageRM stopped
Running vprobed stop
watchdog-vprobed: Terminating watchdog process with PID 3266836
vprobed stopped
Running hostd stop
hostd is not running.
Running lbtd stop
watchdog-net-lbt: Terminating watchdog process with PID 3266811
net-lbt stopped
Running usbarbitrator stop
watchdog-usbarbitrator: Terminating watchdog process with PID 3266789
usbarbitrator stopped
Running SSH stop
SSH login disabled
VobUserLib_Init failed with -1
Running DCUI stop
Disabling DCUI logins
VobUserLib_Init failed with -1
Running DCUI restart
Enabling DCUI login: runlevel =
VobUserLib_Init failed with -1
Running SSH restart
SSH login enabled
VobUserLib_Init failed with -1
Running usbarbitrator restart
Errors:
Not all VMFS volumes were updated; the error encountered was 'IO was aborted by VMFS via a virt-reset on the device'.
Errors:
Not all VMFS volumes were updated; the error encountered was 'IO was aborted by VMFS via a virt-reset on the device'.
Errors:
Rescan complete, however some dead paths were not removed because they were in use by the system. Please use the 'storage core device world list' command to see the VMkernel worlds still using these paths.
Error while scanning interfaces, unable to continue. Error was Not all VMFS volumes were updated; the error encountered was 'No connection'.
usbarbitrator started
Running lbtd restart
net-lbt started
Running hostd restart
[17791] Begin 'hostd ++min=0,swap,group=hostd /etc/vmware/hostd/config.xml', min-uptime = 60, max-quick-failures = 1, max-total-failures = 1000000, bg_pid_file = ''
hostd started.
Running vprobed restart
vprobed started
Running storageRM restart
storageRM started
Running sensord restart
sensord started
Running slpd restart
Starting slpd
Running memscrubd restart
The checkPages boot option is FALSE, hence memscrubd could not be started.
Running dcbd restart
dcbd started
Running cdp restart
cdp started
Running vobd restart
vobd started
Running vpxa restart
[3270235] Begin '/usr/lib/vmware/vpxa/bin/vpxa ++min=0,swap,group=vpxa -D /etc/vmware/vpxa', min-uptime = 60, max-quick-failures = 1, max-total-failures = 1000000, bg_pid_file = ''
Running sfcbd-watchdog restart
Running wsman restart
Starting openwsmand
Running sfcbd restart
This operation is not supported.
Please use /etc/init.d/sfcbd-watchdog start
Hope that helps.
Hi ,
Alt+f12 live kernal logs is more than enough to confirm the same.You have a Storage/Switch issue in your environment.Kernal is not able to detect/query the status of lun's(I cannot confirm how many luns are part of this problem,but looking at the logs lun:naa which is ending with 6b204 is having a problem).Chances are there you may have few dead lun's!!!
Why you are not able to connect to host?
Kernel will be querying the status of all the lun's,and if the commands(SCSI) are failing ,hostd agent will hung after a certain point of time.For the same reason host will get disconnected from VC and Vsphere client to the session will fail(Because hostd agent hung)
There are slight chances you will be able to recover from the same by doing a agent restart.But in this case agent restart is not working(Thats why i asked you how did you confirmed agent restart was succesfull)
Only solution for this problem is restarting your host!!!! Do capture the current stats and check with your Storage team to konw what went wrong with those lun's ? Why the host was not able to perform I/O on those lun's .Hope you are in a better position to move forward
Regard's
Sree
EMCIE|EMCISA-V2|EMCISA|VCP4|VCP5|MCP|MCTS AD2008
Hi,
I hard rebooted the server and everything went back to normal.
Thank you Sreec for your time.
Hi LuteceAdmin,
Glad to hear that
Regard's
Sree
EMCIE|EMCISA-V2|EMCISA|VCP4|VCP5|MCP|MCTS AD2008