cchen2's Posts

I am testing CSE 4.0 for my customer. Does the TKG Cluster created by CSE 4.0 have node auto-healing ability? (the one offered by cluster API, when nodes is done, new node can be automatically deplo... See more...
I am testing CSE 4.0 for my customer. Does the TKG Cluster created by CSE 4.0 have node auto-healing ability? (the one offered by cluster API, when nodes is done, new node can be automatically deployed and merge into the cluster) I understand that the cluster is CAPVCD self-managed. As a feature of CAPVCD, the node auto-healing should work. However, I tried to stop worker node / control plane node, found that nothing happened (No new nodes are created, the nodes stay in "unavailable" status. (I also tried to restart "RDEprojector", and still nothing happened) I want to know if the node auto-healing feature is supposed to be available for cluster created by CSE 4.0
I've successfully deployed a cluster with 2 worker node pools (worker-pool-1 & worker-pool-2) From a k8s user perspective, when I want to deploy a pod to a specified worker pool (e.x. worker-pool-2)... See more...
I've successfully deployed a cluster with 2 worker node pools (worker-pool-1 & worker-pool-2) From a k8s user perspective, when I want to deploy a pod to a specified worker pool (e.x. worker-pool-2), how should I write the yaml? For example, in case of GKE, user can use nodeSelector to specify the node pool like below:   nodeSelector:     cloud.google.com/gke-nodepool: POOL_NAME And the gke-nodepool label is automatically added during the creation. However, when checking the label of TKG cluster created by CSE 4.0, I didn't find any node pool related  labels. Of course users can label by themself, but it may be easier to use if the node pool label can be automatically added. administrator@client:~$ k describe node cse4-cluster-01-worker-pool-2-5694c75b78-kjr9l Name: cse4-cluster-01-worker-pool-2-5694c75b78-kjr9l Roles: <none> Labels: beta.kubernetes.io/arch=amd64 beta.kubernetes.io/os=linux kubernetes.io/arch=amd64 kubernetes.io/hostname=cse4-cluster-01-worker-pool-2-5694c75b78-kjr9l kubernetes.io/os=linux   
I've successfully deployed cluster via CSE 4.0 and tried to upgrade from 1.21.8 to 1.22.9. After submitting the upgrade request via GUI, the upgrade didn't kick off. Then I restarted the rdeprojecto... See more...
I've successfully deployed cluster via CSE 4.0 and tried to upgrade from 1.21.8 to 1.22.9. After submitting the upgrade request via GUI, the upgrade didn't kick off. Then I restarted the rdeprojector pod, and the upgrade process started. However, after waiting for over 40 minutes, I found that although the version of control nodes was successfully upgraded to 1.22.9, but the worker nodes stuck in 1.21.8 and an additional worker node was added (before upgrade 2 worker nodes, now 3 worker nodes) and stuck in process status (checking in cluster confiug api) I suppose that the upgrade process stuck, and want to know what may be the trigger and how to fix. ** Some Advice ** 1. It will be better if GUI can show the progress of the entire upgrade process, or at least show if the upgrade process is finished or under processing. In current version, it is confusing and hard to tell. 2. I noticed that the rolling update is done with maxSurge > 0, it means that cluster API will create more temp nodes for updating. For those resource sensitive tenant, it may be better to offer an option to config the maxSurge manually. 
Hi Niandrew, thanks for your help! After deleting the existing vapp templates and catalogs owned by the user org, the workflow goes well!
And here is the post &response for the RDE instance.  It got a 200 OK, but the urn of the item seems weird.. Instead of "urn:xxx:xxx" it shows like "urn%3Axxx%3Axxx"
I found that the cse process is running, but it seems waiting for the RDEs. And in the cse.log, only the following line is output "querying list of RDEs for processing"
I finished all workloads described in the provider workflow docs. When trying to deploy the a tkg cluster via GUI, I found the error on the TKG ova template page. The template catalog is published ... See more...
I finished all workloads described in the provider workflow docs. When trying to deploy the a tkg cluster via GUI, I found the error on the TKG ova template page. The template catalog is published and shared with the user tenant. I also tried reboot cse, but still not work