All Posts

Hi, thank you for the reply. I am aware they could be selected. My questions are: Has it been tested by VMware? (I do not want to spend time to create a lab to test it, if this scenario has not eve... See more...
Hi, thank you for the reply. I am aware they could be selected. My questions are: Has it been tested by VMware? (I do not want to spend time to create a lab to test it, if this scenario has not even been tested by VMware) Is it supported? (or more exactly will it be a supported configuration when CSE 4 will be in GA.)
I was trying to generate a template but it didn't work testtenant@desktop:~$ clusterctl generate cluster test Error: failed to read "cluster-template.yaml" from provider's repository "infrastructure-... See more...
I was trying to generate a template but it didn't work testtenant@desktop:~$ clusterctl generate cluster test Error: failed to read "cluster-template.yaml" from provider's repository "infrastructure-vcd": failed to get GitHub release v1.0.0: failed to read release "v1.0.0": GET https://api.github.com/repos/vmware/cluster-api-provider-cloud-director/releases/tags/v1.0.0: 404 Not Found [] Clusterctl support for Cloud Director is enabled only on clusterctl 1.2. However, our latest CAPVCD (Cluster API provider for Cloud Director) is still configured to work with Core CAPI 1.1.3 and Cluster 1.1.3 only. We will very soon qualify CAPVCD with CAPI 1.2. We will keep you posted on this.
Hi  If your goal is to automate everything (create, resize, upgrade, delete) via VCD API, that is doable, but you must take extra care to craft the payload correctly.  If your goal is to creat... See more...
Hi  If your goal is to automate everything (create, resize, upgrade, delete) via VCD API, that is doable, but you must take extra care to craft the payload correctly.  If your goal is to create the cluster using CSE 4.0 UI and then use VCD API to manage it further, that is also doable.  If your goal is to create the cluster using CSE 4.0 UI and use kubectl (Cluster API) to manage it further, that is not straightforward and we do not recommend this. There is a way to get this option also working. If this is what you want, let us know. We can provide the details. Which one of the above is your requirement?
This will be available as part of CSE4. The `controlPlaneIP` field of the capi yaml can be edited if capvcd is to be used directly. However that will be part of CAPVCD 1.0.0. (The change is in the ma... See more...
This will be available as part of CSE4. The `controlPlaneIP` field of the capi yaml can be edited if capvcd is to be used directly. However that will be part of CAPVCD 1.0.0. (The change is in the main branch.)
Hello, data center group networks should be able to be selected during cluster creation as long as those group networks are visible to the logged-in VCD user.
I have done some tests regarding this topic with a cluster created with 3 master. If one control plane node is shutdown from vCenter, "get pods -A" continue to work. (As expected) If two control pl... See more...
I have done some tests regarding this topic with a cluster created with 3 master. If one control plane node is shutdown from vCenter, "get pods -A" continue to work. (As expected) If two control plane nodes are shutdown, "get pods -A" doesn't work anymore (Expected) After restarting one of the control plane node "get pods -A" works again, (Expected) So the basic functionality of a multi control plane nodes is working. One issue is that no errors are reported in the events or in status of the cluster from CSE plugin. (Status is "ready") The only thing visible is at load balancer level  which shows that some endpoints are down and VAPP that is noticing some VMs down. Would it be possible to add some kinds of "health" in the CSE plugin? (like all control planes node up and running / worker nodes up and running, load balancer associated to management IP deployed etc) Second issue, I have deleted on purpose one of the control plane VM. As mentioned above no information are reported from the CSE plugin, it still show "3 nodes". It doesn't recreate the missing node (no "auto-heal" , which would be the best) Is there a procedure on how to replace a failed node in such case?
Hi, Now, when exposing a service via load balancer from kubernetes it will pick the first IP address available in the first "IP Allocations" with IP available of the associated edge. It would help ... See more...
Hi, Now, when exposing a service via load balancer from kubernetes it will pick the first IP address available in the first "IP Allocations" with IP available of the associated edge. It would help in disaster recovery scenario to be able to define which IP will be used. Similar to: Use static IP with load balancer - Azure Kubernetes Service | Microsoft Docs Is is already planned to provide such capability in Tanzu CSE? If yes, when will it be available? If not, this is a feature request.
In our VCD event logs we see this message every minute: User 'clusterauthor' (83137b58-3506-416b-a31d-e68962bce07b) authorize together with this event: OAuth token created for client ${oauthToken... See more...
In our VCD event logs we see this message every minute: User 'clusterauthor' (83137b58-3506-416b-a31d-e68962bce07b) authorize together with this event: OAuth token created for client ${oauthToken.clientName}(${oauthToken.clientId}) Event Id: urn:vcloud:audit:f06957fc-5c8d-4dc6-8747-af8d88181f86 Type: token   This resulted in the VCD database filling up the disk which resulted in VCD stopping.   Are other having the same issue, or is this a known issue?
PS. Sent on behalf of @marko_uskokovic 
Yes, it does, but only If I login to each site separately. If I login to rs-bg-2 and switch to rs-bg-1 site from the "Data centers" page of tenant portal, then I get "Error: Failed to fetch Kubernet... See more...
Yes, it does, but only If I login to each site separately. If I login to rs-bg-2 and switch to rs-bg-1 site from the "Data centers" page of tenant portal, then I get "Error: Failed to fetch Kubernetes clusters" because of "401 Unauthorized" request to /api/session and the rezulting "CSE UI: Error fetching sites". Since this is evident even on CSE 3, should I open SR and debug it there and give VMware with screen sharing session so someone can better describe problem and collect what is going on?
Hi @marko_uskokovic  Sorry for the delay; my message did not go through.   Would you please log into each site and go to the container plugin on each site? Does each site show clusters successfull... See more...
Hi @marko_uskokovic  Sorry for the delay; my message did not go through.   Would you please log into each site and go to the container plugin on each site? Does each site show clusters successfully?   Thank you
Hi, I do not have full details but from what i understood: NSX ALB communicate with vCenter using a "vCenter account" dedicated for this purpose. (This is part of "create NSX-T Cloud) in vcenter. ... See more...
Hi, I do not have full details but from what i understood: NSX ALB communicate with vCenter using a "vCenter account" dedicated for this purpose. (This is part of "create NSX-T Cloud) in vcenter. So it seems somehow that NSX-ALB was not able to communicate with vCenter anymore. So maybe password has been modified or something like this. Note: I am maybe mistaken an issue was with account connecting to NSX-Manager but the concept is the same, issue with credentials used with NSX-T cloud) After fixing credentials the deployment was successful. Summary: The issue was not related to Tanzu/CSE but the underlying NSX-ALB infrastructure. Unfortunately it is not easy to pinpoint the origin when looking at error at Tanzu/CSE level. Therefore, the feature requests of adding "pre-requisite" check and/or a wizard showing the progression of a cluster deployment step by step. (Showing the steps completed, current step, and next steps.) In this way it would be easier to pinpoint the origin of such issue if one step is stuck.
Hi, any update on this request? I was trying to generate a template but it didn't work testtenant@desktop:~$ clusterctl generate cluster test Error: failed to read "cluster-template.yaml" from p... See more...
Hi, any update on this request? I was trying to generate a template but it didn't work testtenant@desktop:~$ clusterctl generate cluster test Error: failed to read "cluster-template.yaml" from provider's repository "infrastructure-vcd": failed to get GitHub release v1.0.0: failed to read release "v1.0.0": GET https://api.github.com/repos/vmware/cluster-api-provider-cloud-director/releases/tags/v1.0.0: 404 Not Found [] Or is clusterctl generate still not supported for vcloud director? Like in this link clusterctl generate command doesn't support the generation of CAPI yaml for Cloud Director; Follow the guidelines provided below configure the CAPI Yaml file cluster-api-provider-cloud-director/WORKLOAD_CLUSTER.md at main · vmware/cluster-api-provider-cloud-director · GitHub
Hi, so far I have only deployed the legacy and beta CSE clusters in "Routed Organization Virtual Data Center Network". It works. Is is supported/tested to deploy in a "Routed Data Center Group Netw... See more...
Hi, so far I have only deployed the legacy and beta CSE clusters in "Routed Organization Virtual Data Center Network". It works. Is is supported/tested to deploy in a "Routed Data Center Group Network"? Goal is to see if distributed firewall could be used to isolate clusters deployed in different networks.
I don't see any specific error, because it opens in new tab without inspect element so I can't track http requests. Will report if I find some error. When I login to rs-bg-2-ec site and want to go t... See more...
I don't see any specific error, because it opens in new tab without inspect element so I can't track http requests. Will report if I find some error. When I login to rs-bg-2-ec site and want to go to another site by SSO I just click on site name in "Data Centers" -> "Virtual Data Center" like in picture attached  
Hi @marko_uskokovic, Thank you for your feedback. We are looking into this. When SSO login fails after clicking the cse4-rs-bg-1-test-1 link on the rs-bg-2-ec, do you see any specific error? Wou... See more...
Hi @marko_uskokovic, Thank you for your feedback. We are looking into this. When SSO login fails after clicking the cse4-rs-bg-1-test-1 link on the rs-bg-2-ec, do you see any specific error? Would you please also elaborate on when you log into the rs-bg-2-ec site and go to another site by SSO, what specific steps do you do to get to the other site?   Thank you,   Timothy
>Unfortunately I couldn't easily get into the VM to investigate further - the password found by editing the guest properties in vCD does >not work and nor does connecting from the machine that has th... See more...
>Unfortunately I couldn't easily get into the VM to investigate further - the password found by editing the guest properties in vCD does >not work and nor does connecting from the machine that has the SSH key. Unless a different user (not root) should be used? I have also >reset the password by booting to single-user mode, but it appears to be overwritten by cloud-init once the VM boots again. These are known issues that are getting resolved in GA release > When I restarted the CSE service it deleted the VMs New flag to keep the vm(s) on failure is part of GA. But did the retry on cluster creation happen ?
Thanks for the feedback. This is something we will check with the team. 
I am testing CSE 4.0 for my customer. Does the TKG Cluster created by CSE 4.0 have node auto-healing ability? (the one offered by cluster API, when nodes is done, new node can be automatically deplo... See more...
I am testing CSE 4.0 for my customer. Does the TKG Cluster created by CSE 4.0 have node auto-healing ability? (the one offered by cluster API, when nodes is done, new node can be automatically deployed and merge into the cluster) I understand that the cluster is CAPVCD self-managed. As a feature of CAPVCD, the node auto-healing should work. However, I tried to stop worker node / control plane node, found that nothing happened (No new nodes are created, the nodes stay in "unavailable" status. (I also tried to restart "RDEprojector", and still nothing happened) I want to know if the node auto-healing feature is supposed to be available for cluster created by CSE 4.0
The default wait time for CreateKindVM is 180 seconds, if the time taken more than that, the creation of cluster will be blocked or panic. In my env (a cassini testbed), the process of CreateKindVM t... See more...
The default wait time for CreateKindVM is 180 seconds, if the time taken more than that, the creation of cluster will be blocked or panic. In my env (a cassini testbed), the process of CreateKindVM took about 480 seconds. To make it work, I have to modify the source code to increase the default value and build a temp version to replace the binary in the CSE server VM. The same issue for ClusterCreatePostCustomizationPhases of "guestinfo.cloudinit.target.cluster.ready.status", the default wait time is 1800 seconds, but in my env, it could take more than that time - it took about 1500 seconds in one run, and took more than 1800 seconds in another run.   Can you increase the default values for these two items or make them configurable?