I have two DL385 G7's and for iSCSI connectivity on ESXi 5, I am using two of the internal Broadcom 5709 NIC's with hardware initiator. On my existing ESXi 4.1 hosts, I am using the software iSCSI adapter (NC360T's) which easily connects to an HP p2000 iscsi array and also connects to both Enhance-tech iscsi arrays (no CHAP).
For some reason the 5709 NIC's in the new servers will NOT connect to the enhance-tech SAN. I have upgraded to the latest HP firmware and the latest ESXi 5 build (Dec 15th build 515841)
I get the below error in the vmware syslog.log. Anyone have any ideas? (I can post more of the log if needed)
2012-01-06T23:08:46Z iscsid: Login Target: iqn.1997-10.com.enhance-tech:storage02:target01 if=bnx2i-984be10a54f6@vmk1 addr=10.1.10.5:3260 (TPGT:65536 ISID:0x1)
2012-01-06T23:08:46Z iscsid: Notice: Assigned (H34 T1 C0 session=15, target=5/9)
2012-01-06T23:08:46Z iscsid: SessionResolve for 10.1.10.5 (via vmk1) started)
2012-01-06T23:08:46Z iscsid: Login Target: iqn.1997-10.com.enhance-tech:storage02:target01 if=bnx2i-984be10a54f6@vmk1 addr=10.1.10.6:3260 (TPGT:65536 ISID:0x2)
2012-01-06T23:08:46Z iscsid: Notice: Assigned (H34 T1 C1 session=16, target=5/a)
2012-01-06T23:08:46Z iscsid: SessionResolve for 10.1.10.6 (via vmk1) started)
2012-01-06T23:08:46Z iscsid: DISCOVERY: transport_name=bnx2i-984be10a54f6 Pending=2 Failed=0
2012-01-06T23:08:47Z iscsid: DISCOVERY: transport_name=bnx2i-984be10a54f6 Pending=2 Failed=0
2012-01-06T23:08:48Z iscsid: DISCOVERY: transport_name=bnx2i-984be10a54f6 Pending=2 Failed=0
2012-01-06T23:08:49Z iscsid: DISCOVERY: transport_name=bnx2i-984be10a54f6 Pending=2 Failed=0
2012-01-06T23:08:50Z iscsid: DISCOVERY: transport_name=bnx2i-984be10a54f6 Pending=2 Failed=0
2012-01-06T23:08:50Z iscsid: ep_poll failed rc-1
2012-01-06T23:08:50Z iscsid: Login Failed: iqn.1997-10.com.enhance-tech:storage02:target01 if=bnx2i-984be10a54f6@vmk1 addr=10.1.10.5:3260 (TPGT:65536 ISID:0x1) Reason: 00040000 (Initiator Connection Failure)
2012-01-06T23:08:50Z iscsid: Notice: Reclaimed Channel (H34 T1 C0 oid=5)
2012-01-06T23:08:50Z iscsid: ep_poll failed rc-1
2012-01-06T23:08:50Z iscsid: Login Failed: iqn.1997-10.com.enhance-tech:storage02:target01 if=bnx2i-984be10a54f6@vmk1 addr=10.1.10.6:3260 (TPGT:65536 ISID:0x2) Reason: 00040000 (Initiator Connection Failure)
2012-01-06T23:08:50Z iscsid: Notice: Reclaimed Channel (H34 T1 C1 oid=5)
2012-01-06T23:08:50Z iscsid: Notice: Reclaimed Target (H34 T1 oid=5)
2012-01-06T23:08:51Z iscsid: DISCOVERY: transport_name=bnx2i-984be10a54f6 Pending=0 Failed=2
Which exact model is? (this is an extract from vmware hcl page)
|Broadcom||Broadcom NetXtreme II BCM5709 1000Base-T||iSCSI|
|Broadcom||Broadcom NetXtreme II BCM5709S 1000Base-SX||iSCSI|
I don't know the answer to your question but doing some search, I found this (sadly the web page to this KB seems not to work properly...)
From this page, this is the "interesting" part http://itblog.rogerlund.net/
This issue occurs because ESXi 5.0 attempts to connect to all configured or known targets from all configured software iSCSI portals. If a connection fails, ESXi 5.0 retries the connection 9 times. This can lead to a lengthy iSCSI discovery process, which increases the amount of time it takes to boot an ESXi 5.0 host.
To minimize the amount of time the boot process spends discovering iSCSI targets, you can reduce the number of network portals and the number of targets.
To list the current number and configuration of an ESX host's network portals, run the command:
esxcli iscsi networkportal list
The output is similar to:
MAC Address: 00:1b:21:59:16:e8
MAC Address Valid: true
IPv4 Subnet Mask: 255.255.255.0
Vlan Supported: true
Vlan ID: 10
Reserved Ports: 63488~65536
TCP Checksum: false
Link Up: true
Current Speed: 10000
Rx Packets: 656558
Tx Packets: 111264
NIC Driver: ixgbe
NIC Driver Version: 184.108.40.206.2-10vmw-NAPI
NIC Firmware Version: 0.9-3
Compliant Status: compliant
VswitchUuid: 26 46 30 50 c0 cf df 1e-52 ef ab d7 a2 ab 96 f9
Path Status: active
Note: This is an example of one network portal (HBA34).
To list currently running targets, run the command:
vmkiscsi-tool -T vmhba##
For more information on reducing number of network portals and the number of targets, contact your array vendor”
Hope this throw some light to your problem!!!
Seb - elgreco81
Thanks for your input. The NIC is the Broadcom NetXtreme II BCM5709 1000Base-T, and the wierd thing is ... if I use the iSCSI software intiator all LUN's show up fine. It's only when using the hardware initiator that I get nothing (other than errors in the log). I may need to open a ticket with VMware.
Nothing from them yet. I actually have a HP P2000 G3 as well and the hardware initiator works fine with that unit. I ended up having to use the software initiator for the enhance-tech unit because I couldn't wait any longer. So I'm using both now ugh.
I've had the same issue. What a surprise I am running HP ProLiant DL385 G7 with a P2000 G3. I'm actually dealing with some freezing issues that I am trying to resolve as well. Going to switch to the iSCSI Software Adapter and see if that helps.
I was using both for a few weeks but support told me using both hardware and software initiator simultaneously was not supported. Last week I switched totally over to the software initiator. It seems to work fine but it would be nice to utilize the offload capabilities.
By the way JRedmond, are you using Jumbo Frames? I have not enabled it yet but thinking about it.
I do not have Jumbo Frames enabled. From my knowledge it is not worth using Jumbo Frames unless you are utilitizing 10Gb iSCSI. This is from what I have read and from HP engineers that I've worked with in the past. I wonder if people with an Intel processor G7 have the same issue or it is with the P2000 G3 and ToE in general.
After all of my tests i found a good workaround....seems to be an esxi iscsi bug and the vmware support don't give an official answer.
Try to make 2 vswitch, one for each vmk port.
Configure in the software iscsi adapter all the ip of the host port of HP P2000 g3 in the dynamic discovery section.
Don't bind the vmk port....it do automatically!
It's not the best practice....but the best practice don't work!!!!
Performance is ok and no errors are logged.
look also here:
There is defiantly a bug connecting to an iSCSI array on a second subnet but I found a solution that fits my needs. Maybe it will help.
I was able to ping the iSCSI array from both test host. I was able to 'nc' to port 3260. The host logs showed error 4, transportation error.
I update all firmware on all servers to the most current. It fixed some performance and boot issue but not the iSCSI issue.
Updated the hosts to ESXi 5.0 U2 and installed the recommended NIC drivers. Issue persisted.
Moved the servers to the same subnet and iSCSI worked. Moved back and issue came back.
Installed Windows and Linux guest on the hosts and setup iSCSI initiator software. I was able to connect through the host to and iSCSI target on a different subnet but I couldn’t connect from the host itself.
Tested with a different server and ESXi 5.0 build. Issue persisted.
To resolve the issue, I turned off NIC teaming/failback for all of the configurations sets in the vSwitch that is going to communicate with the iSCSI array. Not just the switch but the VMkernel too.
Next, I made sure that no NIC/vSwitches where bound to the iSCSI software adapter.
This resolved my issue. I tested vMotion, failover, reboot, etc. and everything is working well.
Reading this thread, this won’t work for some of the more complicated setups but for simple setups, this will hopefully resolve your connection issues.