All Posts

HI, we are encountering the same problem, may we know how to resolve this?
Hello, Based on your inputs, I believe the below will help you: Deploying NSX-T in a Stretched Cluster – Part 1  Deploying NSX-T in a Stretched Cluster – Part 2   
you are correct @kastlr. FT isn't support across sites with vSAN Stretched.
The design which the documentation refers to is this: BGP North-South Routing for VMware Cloud Foundation Instances with Multiple Availability Zones In short: You'll have two Edge nodes running in A... See more...
The design which the documentation refers to is this: BGP North-South Routing for VMware Cloud Foundation Instances with Multiple Availability Zones In short: You'll have two Edge nodes running in AZ1 which are protected with vSphere HA and in case of a failure would failover to AZ2. They peer with a routing device in AZ1 and AZ2 respectively and have traffic steering enabled so that e- and ingress is in AZ1 primarily.   From network perspective this is a bad design and leads into long service interruption in case AZ1 fails. I would rather expand this cluster with a second pair of Edge nodes, keep them in AZ2 and peer it accordingly without traffic steering enabled. You'll end up having asymmetric routing but advantages of having almost zero failover time because you'll have active components in every AZ running (no need for failing over Edge nodes between AZ).
I assume you either have a stretched cluster or two site local clusters in VCF but in both cases your intention is to span a Tier-0 GW across the two sites. For redundancy and performance reasons I ... See more...
I assume you either have a stretched cluster or two site local clusters in VCF but in both cases your intention is to span a Tier-0 GW across the two sites. For redundancy and performance reasons I would create an Edge cluster with 4 Edge nodes total, two per site. Depending if your uplink VLANs are stretched across both sites or if they are only available within each site: - All Edges in AZ1 and AZ2, in essence the Tier-0 GW, have their uplink interfaces in the two stretched uplink VLANs. All Edge nodes peer with a router in AZ1 and AZ2 through both uplink VLANs, 4 peers in total. - Each pair of Edge nodes, in essence the Tier-0 GW, have their uplink interfaces in the two site local uplink VLANs in AZ1 and AZ2 respectively. Edge nodes in AZ1 peer with a router in AZ1 and Edge nodes in AZ2 respectively, 4 peers in total. For both cases: Depending on how you wan't your traffic carried to your physical network, you might use traffic steering options in BGP like local preferences and AS path prepending to have e- and ingress traffic though AZ1 primarily. Or you don't steer your traffic but you'll have asymmetric routing (which doesn't need to be negative).
Hi,   what you're asking for is named Fault Tolerance VMs. AFAIK it's not supported in combination with vSAN stretched Cluster, but I'm happy to get corrected. As an Fault Tolerance protected VM ... See more...
Hi,   what you're asking for is named Fault Tolerance VMs. AFAIK it's not supported in combination with vSAN stretched Cluster, but I'm happy to get corrected. As an Fault Tolerance protected VM requires another "shadow" VM which is in sync (Files & Disk, RAM Content, CPU Registers...) that feature does really require a superfast and low latency link between the nodes which would host those paired VMs.
Create DRS should-rules to keep VMs running on the Preferred site to avoid them restarting when taking down the Secondary site.
Hello Guys We have configured Stretched Cluster, which we have tested to turn off an entire datacenter and the VMs were moved to the Secondary Site but at the time of testing the VMs were restarted ... See more...
Hello Guys We have configured Stretched Cluster, which we have tested to turn off an entire datacenter and the VMs were moved to the Secondary Site but at the time of testing the VMs were restarted and it took two minutes to complete the entire process. I have a question if it is possible to have zero downtime of the VMs if one of the sites goes down and the VMs must be moved to the active site, I will appreciate if you can clarify the doubt Thank you
Good night Could you please clarify the following question for me? When Edge Clusters are created in VCF, a minimum of two 2 edges must be added but it is not clear to me if one Edge should go with... See more...
Good night Could you please clarify the following question for me? When Edge Clusters are created in VCF, a minimum of two 2 edges must be added but it is not clear to me if one Edge should go with the BGP values of DC1 and the second Edge with the BGP values of DC2. The above is correct, now I can create 4 Edges (2 for each Datacenter) The parameters I have are the following: BGP NSX-T: 65001 BGP Switch Peer ASN: 65101 (ToR AZ1) Primary Tier 0 Uplink Interface: 192.168.47.2-3/24 BGP Peer IP: 192.168.47.254 Secondary Tier 0 Uplink Interface: 192.168.48.2-3/24 BGP Peer IP: 192.168.48.254 BGP Switch Peer ASN: 65102 (Tor AZ2) Primary Tier 0 Uplink Interface: 192.168.47.130-131/24 BGP Peer IP: 192.168.47.253 Secondary Tier 0 Uplink Interface: 192.168.48.130-131/24 BGP Peer IP: 192.168.48.253 Thanks for your comments Best Regards SAN
Hello, Kindly allow me to understand your scenario by answering the below: - VCF is stretched from VSAN perspective so far correct? Yeah that's right - What is your use case from networking persp... See more...
Hello, Kindly allow me to understand your scenario by answering the below: - VCF is stretched from VSAN perspective so far correct? Yeah that's right - What is your use case from networking perspective between the two sites? A/A ? A/P ? Activate/Activate - What is your expectation about the failure scenario from Site A to Site B from time perspective? Let the VMs move and be able to maintain IP addressing   Regards
Hello, kindly allow me to understand your scenario by answering the below: - VCF is stretched from VSAN perspective so far correct ? - What is your use case  from networking perspective between th... See more...
Hello, kindly allow me to understand your scenario by answering the below: - VCF is stretched from VSAN perspective so far correct ? - What is your use case  from networking perspective between the two sites ? A/A ? A/P ? - What is your expectation about the failure scenario from Site A to Site B from time perspective ?  
Hello Guys Sorry for the question, because I have confusion regarding Configure BGP in the Tier-0 Gateway for Availability Zone 2. I have configured VCF and it is already stretched with two datacen... See more...
Hello Guys Sorry for the question, because I have confusion regarding Configure BGP in the Tier-0 Gateway for Availability Zone 2. I have configured VCF and it is already stretched with two datacenters, which are operational. Now I am going through the process of Configure BGP in the Tier-0 Gateway for Availability Zone 2, but I have doubts in the process indicated in https://docs.vmware.com/en/VMware-Cloud-Foundation/5.0/vcf-admin/GUID-E2A6DCE6-EC70-4583-AB45-5504C7850651.html   On the NSX Manager main navigation bar, click Networking. In the navigation pane, click Tier-0 gateways. Select the gateway and from the ellipsis menu, click Edit. Add the uplink interfaces to the NSX Edge nodes.             Expand BGP and in the BGP neighbors section, click 2.             In the Set BGP neighbors dialog box, click Add BGP neighbor and configure the following settings In Sources address, it asks me for the interfaces of AZ2, but it only shows me the interfaces of AZ1 IPs: 192.168.47.2 192.168.47.3 192.168.48.2 192.168.48.3 NSX ASN is 65000 ASN Switch AZ1: 65101 ASN Switch AZ2: 65102 192.168.47.130 192.168.47.131 192.168.48.130 192.168.48.131 I don't know if it's correct, or I should do some process before. Thanks for your comments
This is not relevant. We are not talking about the vMotion network here.
Recommendation To test the MTU size of 9,000, the database was running an OLTP workload while the virtual machine was moved using vMotion to another server. We captured the time to vMotion the serve... See more...
Recommendation To test the MTU size of 9,000, the database was running an OLTP workload while the virtual machine was moved using vMotion to another server. We captured the time to vMotion the server as well as New Orders per Minute (NOPM) and Transactions per Minute (TPM) for analysis. Increasing the MTU size showed significant gains for these metrics: vMotion time to complete NOPM TPM A larger MTU size of 9,000 reduced the time to complete moving the database with an active workload compared to the default MTU size of 1,500. This impressive time-saving function led us to label this best practice as a Day 1, Highly Recommended procedure.  NOPM and TPM also increased, but not as substantially as the vMotion time savings. The overall benefit of this best practice is that database performance has greater consistency during vMotion with a substantial time-savings in moving the database.  Implementation Steps Process to configure MTU size = 9000 for vMotion network Configure MTU Size = 9000 at vMotion Distributed Switch To change at vMotion distributed switch level, select the distributed switch for vMotion in networking tab in vSphere and select setting In the Edit setting, select Advanced and change the MTU to 9000 Click OK to save the change Configure MTU Size = 9000 at VMKernel Adapter Login to vSphere and select the EXSi host, and click the Configure tab  Select the VMkernel adapters  Under VMkernel adapters, select the Motion network and click Edit In the Edit Settings, under VMkernel port settings, change the value of MTU from 1500 to 9000. Click OK to save the change Repeat the same steps above on another host Configure MTU Size = 9216 at Physical Switch To change at switch level, show running configuration interface ethernet (Run this to find port rang Configure Terminal Interface range ethernet "enter the port range" (eg:1/1/25 -1/1/2 mtu 92 Exit Write memory to save the configuration Note: To change it at network switch level, login into switch console and update the MTU
It should not be disruptive as far as I know.
Thank you for documenting this. I was leaning towards creating a new vlan TZ.  And applying to the nodes using the Transport Node Profile that VCF created.  Is  attaching, or detaching, the TNP to ... See more...
Thank you for documenting this. I was leaning towards creating a new vlan TZ.  And applying to the nodes using the Transport Node Profile that VCF created.  Is  attaching, or detaching, the TNP to a Baseline WLD disruptive?  I assume in this case only adding additional TZ to the nodes, so it won't take VMs or nodes offline.
Here are some steps you can take: NSX-T Load Balancer Configuration: Check the NSX-T load balancer configuration for the vROps cluster. Make sure that the load balancer is properly configured with... See more...
Here are some steps you can take: NSX-T Load Balancer Configuration: Check the NSX-T load balancer configuration for the vROps cluster. Make sure that the load balancer is properly configured with the correct virtual server, pool, and health check settings. Confirm that the pool members (vROps nodes) are added correctly and their status is healthy. Health Checks: Verify that the health check monitors defined in NSX-T are configured properly. The health check monitors should be configured to check the correct services on the vROps nodes, and they should match the services that are configured to listen on the vROps cluster IP. Firewall Rules: Review the firewall rules within NSX-T to ensure that traffic to the vROps cluster IP is allowed. Both inbound and outbound rules might need to be reviewed to ensure proper communication. Network and Routing: Check the network and routing configurations. Ensure that there are no routing issues preventing traffic from reaching the vROps cluster IP. Check if any network segments or routes need to be configured in NSX-T to allow proper communication. DNS Resolution: Confirm that the DNS records for the vROps cluster IP and FQDN are correctly configured and resolving to the appropriate IP addresses. DNS issues can lead to communication problems. vROps Node Health: Verify the health of individual vROps nodes. If one of the nodes is experiencing issues, it might affect the overall cluster accessibility. Check the vROps management console for any alerts or errors. NSX-T Logs and Monitoring: Monitor NSX-T logs and events related to load balancer activity. This can provide insights into any issues that might be occurring at the network level. vROps Logs: Check the vROps logs for any errors or issues that might be related to the cluster IP accessibility. This could provide additional context about what's happening on the vROps side.
Also read through below 7 minutes blog, would give you much more insight on how to use DFW only feature in NSX-T, https://www.vgarethlewis.com/2022/06/20/vmware-nsx-micro-segmentation-only-deploymen... See more...
Also read through below 7 minutes blog, would give you much more insight on how to use DFW only feature in NSX-T, https://www.vgarethlewis.com/2022/06/20/vmware-nsx-micro-segmentation-only-deployment/ Credit to Gareth Lewis -Blog
Hi, you need to attach a VLAN TZ to be able to create VLAN Backed Segments. Take a look at my blog post for details: https://cybernils.net/2023/07/17/creating-vlan-backed-segments-in-vmware-cloud-fo... See more...
Hi, you need to attach a VLAN TZ to be able to create VLAN Backed Segments. Take a look at my blog post for details: https://cybernils.net/2023/07/17/creating-vlan-backed-segments-in-vmware-cloud-foundation-vcf/  
Hello Guys I have installed VRLCM and deployed vROPS, and the master and replica nodes respond without problems and access to vROPS through the fqdn, but the vROPS cluster ip does not respond. I che... See more...
Hello Guys I have installed VRLCM and deployed vROPS, and the master and replica nodes respond without problems and access to vROPS through the fqdn, but the vROPS cluster ip does not respond. I checked in NSX-T and the LB is enabled, but it does not respond to ping or access to the portal. Could someone tell me if I should check something in NSX-T, or what may be happening that the vrops cluster does not respond. Thanks for your help.   Regards SAN