Czernobog's Posts

The issue is resolved for me - when placing the edges on NSX-prepared hosts, the geneve tunnel traffic has to be placed on a separate NSX-segment. So, if someone has the same issue, create a new VLAN... See more...
The issue is resolved for me - when placing the edges on NSX-prepared hosts, the geneve tunnel traffic has to be placed on a separate NSX-segment. So, if someone has the same issue, create a new VLAN NSX Segment, add it to a VLAN TZ, add the TZ to your edges. This way you will be able to select this segment for your tunnel traffic. The NSX design reference guide has just recently been updated to match NSX 3.0 (no 3.1 still...), I hope the business unit steps up in this regard. edit: I have no idea how to mark the post as "resolved" after the recent vmtn update, so I'll just mark my answer as the correct one.
Thank you  for the offer, however this is not something that would be allowed in my environment:) I'll have to wait for GSS to finally respond to my SR.
I have followed another guide, yes, your link reutrns a 404
I've deployed a fresh 3.1 environment, configured my transport nodes, Tier-0, Tier-1 and attached a few segments, but as soon as I add a VM to a segment, the tunnel becomes degraded. I want to run a... See more...
I've deployed a fresh 3.1 environment, configured my transport nodes, Tier-0, Tier-1 and attached a few segments, but as soon as I add a VM to a segment, the tunnel becomes degraded. I want to run a "collapsed" *environment now, where the TEP vmkernels run on the vds too and getting this ocnfiguration to work was my goal with the 3.1 update. I guess the error is somewhere in the transport node (TEP) configuration, which I have to figure out now...
After discussing the issue with GSS we came to the conclusion, that it would just be easier to redeploy the environment. There seemed to be some inconsistency in the corfu db, that was probably caus... See more...
After discussing the issue with GSS we came to the conclusion, that it would just be easier to redeploy the environment. There seemed to be some inconsistency in the corfu db, that was probably caused when the environment was shrinked from 3 to 1 nodes, but there was still a paused (failed) upgrade active. So do not remove or add manager nodes when there is still an active upgrade process present, and you should be fine.
I used copying files to identify the problem. What would be the benefit of using iperf in this case? The MTU is set correctly end to end, 1600 as per documentation. Wouldn't I get dropped packets, w... See more...
I used copying files to identify the problem. What would be the benefit of using iperf in this case? The MTU is set correctly end to end, 1600 as per documentation. Wouldn't I get dropped packets, when the MTU is not set correctly somewhere along the way, sicne Geneve frames cannot be fragmented? Also it does not seem to be an issue with the networking hardware, since the problem occurs even the source and target system are placed on the same ESXi host. Packets would not leave the host in this case.
Here are the 3 profiles applied on the transport nodes (ignore _VDS, it is not applied anywhere). The first one is applied on the transport node n-vds uplinks (vmnic2 and vmnic3). The second one on... See more...
Here are the 3 profiles applied on the transport nodes (ignore _VDS, it is not applied anywhere). The first one is applied on the transport node n-vds uplinks (vmnic2 and vmnic3). The second one on the TEP interface of the edge node (fp-eth0). The third on the uplink ports fp-eth1 and fp-eth2. Edit: I've also just checked, and the transfer rate between overlay networks attached to different edge node clusters is also affected. The other ledge cluster is identically configured as the one described above, just that the logical switch uses another subnet for adressing. Also - AFAIK the transport node TEPs as well as the edge TEPs can use IPs from the same subnet. I could not find contradictory information. I know that when you use NSX-T lower than 3.1 in regards of VLANs in the transport network, but this occurs only when you use the VDS 7 exclusively to channel all traffic. This is not the case here.
The Edges are deployed as VMs. They are grouped in a cluster as a pair, with active-passive failover mode. Here's a scetch of a single node, I hope this helps: edit: added some info.
- have to repost this, since yesterday the new great forum software did not seem to save the thread... - I am managing a small NSX-T test environment, consisting of 3 ESXi hosts as transport nodes a... See more...
- have to repost this, since yesterday the new great forum software did not seem to save the thread... - I am managing a small NSX-T test environment, consisting of 3 ESXi hosts as transport nodes and an overlay transport zone, where a few overlay networks are configured. The hosts have 4 x 1Gbps physical uplinks (no switches with 10gb available atm...), two connected to the vds, the rest to the n-vds. The Uplink profile applied to the ESXi Hosts uses both n-vds uplinks in a LB configuration. It uses the global MTU configuration setting of MTU = 1600. 2 Edge Transport nodes are grouped in an Edge Cluster. On the edges, uplink-1 is connected to the transport vlan for the overlay networks, uplinks 2 and 3 are used for communication with the physical switches. There is almost no load on the environment currently, be it compute, storage or network. The edge nodes are deployed in the Medium sizing configuration. The problem I am facing is, that transfer rates between vms placed on the vds in a physical vlan and VMs placed in the overlay network are extremely bad. I can only get something needing low thorughput, like SSH, to work, but during file transfers using SSH or HTTP I get as low as 30kB/s before loosing the connection. Using performance views from the vcenter I did not notice any packet drops neither on the hosts' vmnics nor the VMs' nic. Ping between components returns roundtrip times of <1 ms. Transfer rates between VMs inside the overlay network are fine, even when they are placed on different ESXi hosts in the cluster. I've also tried different source systems from the physical VLAN to do the transfer tests. All VMs placed in the overlay, regardles of the ESXi host, seem to be affected. Transfers between systems placed in VLANs on the vds are not negatively affected either. All switchports connecting the transport nodes are configured consistenlty in regards of VLAN trunks, MTU size being 9216 bytes. I've used https://spillthensxt.com/how-to-validate-mtu-in-an-nsx-t-environment/ as a guideline to check for MTU consistency across all components and could not find an issue. I use 1600 on the vds and as a default value on the nsx-t uplink profiles used on the transport nodes as well as the edge nodes. I'm kind of at a loss here what to troubleshoot next, and any tips are most welcome:)
I am managing a small NSX-T test environment, consisting of 3 ESXi hosts as transport nodes and an overlay transport zone, where a few overlay networks are configured. The hosts have 4 x 1Gbps physi... See more...
I am managing a small NSX-T test environment, consisting of 3 ESXi hosts as transport nodes and an overlay transport zone, where a few overlay networks are configured. The hosts have 4 x 1Gbps physical uplinks (no switches with 10gb available atm...), two connected to the vds, the rest to the n-vds. The Uplink profile applied to the ESXi Hosts uses both n-vds uplinks in a LB configuration. It uses the global MTU configuration setting of MTU = 1600. 2 Edge Transport nodes are grouped in an Edge Cluster. On the edges, uplink-1 is connected to the transport vlan for the overlay networks, uplinks 2 and 3 are used for communication with the physical switches. There is almost no load on the environment currently, be it compute, storage or network. The edge nodes are deployed in the Medium sizing configuration. The problem I am facing is, that transfer rates between vms placed on the vds in the physical vlan network and VMs placed in the overlay network are extremely bad. I can only get something needing low thorughput, like SSH, to work, but during file transfers using SSH or HTTP I get as low as 30kB/s before loosing the connection. Using performance views from the vcenter I did not notice any packet drops neither on the hosts' vmnics nor the VMs' nic. Ping between components returns roundtrip times of <1 ms. Transfer rates between VMs inside the overlay network are fine, even when they are placed on different ESXi hosts in the cluster. I've also tried different source systems from the physical VLAN to do the transfer tests. All VMs placed in the overlay, regardles of the ESXi host, seem to be affected. Transfers between systems placed in VLANs on the vds are not negatively affected either. All switchports connecting the transport nodes are configured consistenlty in regards of VLAN trunks, MTU size being 9216 bytes. I've used https://spillthensxt.com/how-to-validate-mtu-in-an-nsx-t-environment/ as help and check for MTU consistency acrros all components and could not find an issue. I use 1600 on the vds and as a default value on the nsx-t uplink profiles used on the transport nodes as well as the edge nodes. I'm kind of at a loss here what to troubleshoot next, and any tips are most welcome:)
I have tried it with my LDAP-bound account (Enterprise Admin role) and also the default admin. Tried logging into the applaince by FQDN and IP, also tried to delete all cluster nodes except one, whi... See more...
I have tried it with my LDAP-bound account (Enterprise Admin role) and also the default admin. Tried logging into the applaince by FQDN and IP, also tried to delete all cluster nodes except one, whiel also removing the cluster VIP. Edit: one of the manager nodes cannot sync the repository. Also the removal of a manager node the past, and the resulting leftover information might be the cause of upgrades not working.
I want to do an upgrade to 3.1 When selecting the Upgrade option in the UI, I get a blank screen: Tried different browsers, relogging, restarting the ui service. Any idea what can cause this? ... See more...
I want to do an upgrade to 3.1 When selecting the Upgrade option in the UI, I get a blank screen: Tried different browsers, relogging, restarting the ui service. Any idea what can cause this? edit: all other UI elements are displayed witohut issues. edit2: I currently only have one NSX Manager node. Had more before, due to a hardware migration the cluster was reduced to one node. The upgrade service is running: Service name: install-upgrade Service state: running Enabled on: [ip of the mgr node]
Ah, nice, I've just downloaded the upgrade bundle and will see if it works.
Last week I wanted to migrate workloads from the N-VDS to the VDS7, following a vSphere 7 upgrade. The ESXi hosts in my test environment have 4 NICs each, the initial configuraton was vmnic0 + vmnic... See more...
Last week I wanted to migrate workloads from the N-VDS to the VDS7, following a vSphere 7 upgrade. The ESXi hosts in my test environment have 4 NICs each, the initial configuraton was vmnic0 + vmnic1 on vds, vmnic2 + vmnic3 on n-vds. The migration and new uplink assignment was done successfully, at least I did not find an error in the uplink profile configuration of the transport nodes. First the vmnic0 and vmnic1 were used in the profile, after the new vmkernels were online I have assigned the hosts vmnic2 and vmnic3 to the vds. What immedietally became obvious was, that the overlay tunnel status switched to degraded. Falling back to the n-vds configuration remediated the issue. I have checked my configuration and looked for clues in the documentation, but could not find a solution. Finally I came upon this blog post, where the summary explains the behaviour well: "So when you're using and N-VDS or VDS for NSX-T and you're placing an Edge on the same switch you have to put the Edge overlay in a different subnet. The Geneve traffic that originates from the Edge is not allowed to pass a switch that's hosting a tunnel endpoint for ESXi (VMK10)." Following this advice, I've created a new VDS with only one port group for the overlay traffic, and connected to vmnics of the host to its. After this, the nsx edges nic dedicated to te overlay tunnel connection was attached to this new port group, and the tunnel was re-established. This poses a problem however, because in production I have hosts with only w nics. This means, that I would have to separate the vmnics somehow between two distributed switches, which woudl resultat in a non-redundand setup. Is there a way to separate the overlay and endpoint traffic while still placing all vmkernels on the same VDS?
Thanks, the info helped, I just wanted to be sure. I've applied the NSX-T key (Data Center E+) and there was no negative impact. All features are there and the vCenter still shows 0 CPU usage ... See more...
Thanks, the info helped, I just wanted to be sure. I've applied the NSX-T key (Data Center E+) and there was no negative impact. All features are there and the vCenter still shows 0 CPU usage
I run an environment where NSX-V 6.4.8 is deployed. For some time now I am preparing to move to NSX-T, which a test env running for some time now. The prod environment is constanlty expanding, ... See more...
I run an environment where NSX-V 6.4.8 is deployed. For some time now I am preparing to move to NSX-T, which a test env running for some time now. The prod environment is constanlty expanding, but, since we plan to move to NSX-T at some point, my company hase been purchasing only NSX-T licences, not NSX-V. This leaves the prod environment with NSX-V underlicensed - for the 10 Hosts in my inventory, I have only 10 NSX-V CPU licenses, BUT I have enough NSX-T licenses to cover all the demand. This brings me to the point of this thread - the transition to NSX-T is delayed, but I do not want to leave the prod environment underlicensed. So, I want to apply my NSX-T licensing key to the NSX-V solution in the vCenter, up until I get NSX-T in production. Both NSX-V and NSX-T licenses are from the "Enterprise (Plus)" tier. When trying to do this, I get following message: From my experience, it does not *really* matter what license is applied to the NSX-V solution, hence the underlicensed state - NSX does not seem to mind. But I am afriad this time all of the listed features might be deactivated and I would have apply the NSX-V key again and recover functionality. Can anyone give a tip on how to handle this?
Did you explicitly state in the header, that the content type will be json in your api call? As in content-type = application/json. Per default I think the API expects text/xml.
Here's an issue I have discovered lately: I have configured the DHCP service on one of the NSX Edges. The DHCP is supposed to give out addresses in a /23 network segment, by the use of Bindings.... See more...
Here's an issue I have discovered lately: I have configured the DHCP service on one of the NSX Edges. The DHCP is supposed to give out addresses in a /23 network segment, by the use of Bindings. There is no IP pools configured, the bindings are created automatically when a new VM is created. Example IP range used for the bindings: 10.0.1.0/23, mask 255.255.254.0. The Interface of the Edge is 10.0.1.3, so in the same broadcast domain. The problem I have stumbled upon is, that when the first half of the address range is exhausted, so that all adresses up to 10.0.1.255 are used, VMs that get offers from the range 10.0.2.0-10.0.2.254 do not get the IP assigned. After some troubleshooting I found out that an incorrect subnet mask is handed out by the DHCP server, regardless of what is configured in the binding. In my case 255.255.255.0 is handed out, instead of 255.255.254.0, which is configured in the binding. So when IPs get offered in the 10.0.2.x range, they use the broadcast address 10.0.2.255/24 instead of 10.0.1.255/23 and the DHCP request does not reach the DHCP interface. Is this a known issue? edit: edited for more clarity edit2: It turns out, that the edge had sort of a stale IP Pool configuration. Every time a new VM was taken online, even though there is a binding configured for it, the Edge DHCP Service would try and hand out an address from this IP Pool, which failed because as per log no new leases could be created as the pool was depleted. Also, the DHCP offers that were handed out before included a /24 mask instead of a /23 mask, because of which sending a request would fail anyway, since a wrong boradcast address would be used. Clearing all leases on the edge by nsx cli and also redeploying the edge did not remediate the issue. I had to deply a new edge and migrate the bindings.
AFAIK vSphere 7 compatibility will be built in with NSX 6.4.7, will have to wait for the announcement...