{"tests":[{"name":"cl.ignition.v1.noop","result":"FAIL","duration":82630263197,"output":"        harness.go:646: Cluster failed starting machines: error creating IP address: 504 Gateway Timeout, status code 504, Body: upstream request timeout\n"},{"name":"kubeadm.v1.35.1.flannel.base","result":"FAIL","duration":117044828909,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: error creating IP address: 504 Gateway Timeout, status code 504, Body: upstream request timeout\n"},{"name":"cl.install.cloudinit","result":"FAIL","duration":120399864925,"output":"        harness.go:646: Cluster failed starting machines: error creating IP address: 504 Gateway Timeout, status code 504, Body: upstream request timeout\n"},{"name":"coreos.ignition.once","result":"PASS","duration":592133785608,"output":""},{"name":"cl.ignition.v2.btrfsroot","result":"FAIL","duration":798859136670,"output":"        harness.go:646: Cluster failed starting machines: machine \"06201608-7237-476b-b4cd-a4ae978739be\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.118.19:22: i/o timeout\n"},{"name":"kubeadm.v1.33.8.calico.base","result":"FAIL","duration":91619728939,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: error creating IP address: 504 Gateway Timeout, status code 504, Body: upstream request timeout\n"},{"name":"coreos.ignition.sethostname","result":"FAIL","duration":817502571016,"output":"        harness.go:646: Cluster failed starting machines: machine \"36f80320-44c5-469d-850f-0075393f73da\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.116.123:22: i/o timeout\n"},{"name":"coreos.ignition.security.tls","result":"FAIL","duration":976024935660,"output":"        harness.go:646: Cluster failed starting machines: machine \"85bc6e60-98de-4a45-8dca-7e3e3daf6396\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.66.153:22: i/o timeout\n"},{"name":"coreos.ignition.resource.remote","result":"FAIL","duration":1025876325337,"output":"        harness.go:646: Cluster failed starting machines: machine \"40312796-0e41-47cd-ab19-80f59692d1d8\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.65.136:22: i/o timeout\n"},{"name":"kubeadm.v1.34.4.cilium.base","result":"FAIL","duration":1027351568794,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"835ab3fa-12d9-47eb-bf7f-f5ee1d81a3b9\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.110.6:22: i/o timeout\n"},{"name":"linux.nfs.v4","result":"FAIL","duration":1085038244723,"output":"        nfs.go:77: Cluster.NewMachine: machine \"0d2f9451-531d-4dd3-bcff-11c8f0b7b855\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.64.198:22: i/o timeout\n"},{"name":"kubeadm.v1.34.4.flannel.base","result":"FAIL","duration":1053777631886,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"e6184e8c-9480-4504-bf7d-bff7fead7c66\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.66.167:22: i/o timeout\n"},{"name":"cl.cloudinit.basic","result":"FAIL","duration":1139827002805,"output":"        harness.go:646: Cluster failed starting machines: machine \"fb72b6d5-d9c5-4474-a726-dd8afda08070\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.65.206:22: i/o timeout\n"},{"name":"coreos.ignition.ssh.key","result":"FAIL","duration":1190444102933,"output":"        harness.go:646: Cluster failed starting machines: machine \"8d6d2d80-5dfd-4332-a427-630ee18ba27a\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.67.227:22: i/o timeout\n"},{"name":"cl.flannel.vxlan","result":"FAIL","duration":1136314148905,"output":"        harness.go:646: Cluster failed starting machines: machine \"dfaa3920-dd43-44fe-a058-a5e452d43412\" failed basic checks: ssh unreachable or system not ready: failure checking if machine is running: systemctl is-system-running returned stdout: \"starting\", stderr: \"\", err: Process exited with status 1, systemctl list-jobs returned stdout: \"JOB  UNIT                        TYPE  STATE\\n1075 etcd-member.service         start running\\n875  multi-user.target           start waiting\\n1027 flannel-docker-opts.service start waiting\\n1026 flanneld.service            start waiting\\n\\n4 jobs listed.\", stderr: \"\", err: \u003cnil\u003e\n"},{"name":"cl.ignition.kargs","result":"PASS","duration":218009298092,"output":"        cluster.go:152: + cat /proc/cmdline\n"},{"name":"cl.network.initramfs.second-boot","result":"FAIL","duration":842546292561,"output":"        harness.go:646: Cluster failed starting machines: machine \"d1f7f801-0526-465e-93e3-d64107327bfd\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.85.62:22: i/o timeout\n"},{"name":"docker.network-openbsd-nc","result":"FAIL","duration":886945012138,"output":"        harness.go:646: Cluster failed starting machines: machine \"5867850d-13c6-4bab-86da-cec65add373c\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.84.220:22: i/o timeout\n"},{"name":"kubeadm.v1.33.8.flannel.base/node_readiness","result":"PASS","duration":10782028535,"output":""},{"name":"kubeadm.v1.33.8.flannel.base/nginx_deployment","result":"PASS","duration":11074017122,"output":""},{"name":"kubeadm.v1.33.8.flannel.base/NFS_deployment","result":"PASS","duration":17433628395,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.33.8.flannel.base","result":"PASS","duration":774618111055,"output":"        cluster.go:125: I0429 00:20:15.710343    2157 version.go:261] remote version is much newer: v1.36.0; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0429 00:20:27.258011    2373 version.go:261] remote version is much newer: v1.36.0; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669-0-0-n-bc5a4f368d\" could not be reached\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669-0-0-n-bc5a4f368d\": lookup ci-4669-0-0-n-bc5a4f368d on 1.1.1.1:53: no such host\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [ci-4669-0-0-n-bc5a4f368d kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 10.0.1.65]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.501520853s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.1.65:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.643244928s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 2.129485942s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 4.001402267s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669-0-0-n-bc5a4f368d as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669-0-0-n-bc5a4f368d as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 45ox85.oudp001cixwlerzn\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.1.65:6443 --token 45ox85.oudp001cixwlerzn \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:3bccb1a31b83af0156fea9ddd119b16c6c52d3cdc36a191eab695ff6134fbfd4 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0429 00:23:42.012501    1924 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.flannel.base/node_readiness (10.78s)\n    --- PASS: kubeadm.v1.33.8.flannel.base/nginx_deployment (11.07s)\n    --- PASS: kubeadm.v1.33.8.flannel.base/NFS_deployment (17.43s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"cl.flannel.udp","result":"FAIL","duration":1803418741820,"output":"        harness.go:608: Cluster failed: creating network for cluster: WaitWithContext() has timed out: context deadline exceeded\n"},{"name":"cl.internet","result":"FAIL","duration":1803834989937,"output":"        harness.go:608: Cluster failed: creating network for cluster: WaitWithContext() has timed out: context deadline exceeded\n"},{"name":"kubeadm.v1.34.4.calico.base","result":"FAIL","duration":1803896248251,"output":"        harness.go:608: Cluster failed: creating network for cluster: WaitWithContext() has timed out: context deadline exceeded\n"},{"name":"cl.basic","result":"FAIL","duration":1803896264515,"output":"        harness.go:608: Cluster failed: creating network for cluster: WaitWithContext() has timed out: context deadline exceeded\n"},{"name":"cl.ignition.misc.empty","result":"FAIL","duration":1803917443721,"output":"        harness.go:608: Cluster failed: creating network for cluster: WaitWithContext() has timed out: context deadline exceeded\n"},{"name":"cl.etcd-member.discovery","result":"FAIL","duration":835869602471,"output":"        harness.go:646: Cluster failed starting machines: machine \"6d12a6d0-316d-4f75-89fa-3bffd00aed8c\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.118.250:22: i/o timeout\n"},{"name":"linux.nfs.v3","result":"FAIL","duration":792005169324,"output":"        nfs.go:77: Cluster.NewMachine: machine \"262b4c61-822d-4010-bcb9-0144ddfe8974\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.66.68:22: i/o timeout\n"},{"name":"kubeadm.v1.35.1.cilium.base","result":"FAIL","duration":989816734922,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node: machine \"f70eb779-914e-4058-aa33-10d5215a4d5b\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.65.7:22: i/o timeout\n"},{"name":"kubeadm.v1.33.8.cilium.base","result":"FAIL","duration":1174596802693,"output":"        cluster.go:125: I0429 00:14:50.753627    2200 version.go:261] remote version is much newer: v1.36.0; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0429 00:15:01.590399    2432 version.go:261] remote version is much newer: v1.36.0; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669-0-0-n-d7ff634c54\" could not be reached\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669-0-0-n-d7ff634c54\": lookup ci-4669-0-0-n-d7ff634c54 on 1.1.1.1:53: no such host\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [ci-4669-0-0-n-d7ff634c54 kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 10.0.0.164]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.501469633s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.164:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 2.188951226s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 2.421793482s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 4.002104822s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669-0-0-n-d7ff634c54 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669-0-0-n-d7ff634c54 as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 3v67pr.nhobemod24n9w2qx\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.164:6443 --token 3v67pr.nhobemod24n9w2qx \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:6d3b249e85ffe7d67d41cecf6fc0fc0c6035df426e1d84a5c86681a2f7027716 \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        kubeadm.go:197: unable to setup cluster: unable to create worker node: machine \"205090b2-e277-4bff-9331-16b269be5539\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.67.227:22: i/o timeout\n"},{"name":"cl.ignition.v2.noop","result":"FAIL","duration":772848273982,"output":"        harness.go:646: Cluster failed starting machines: machine \"526c7bfd-98ee-4800-ac0c-d4ec4e96e3ce\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.66.248:22: i/o timeout\n"},{"name":"coreos.ignition.resource.local","result":"FAIL","duration":865932904410,"output":"        harness.go:646: Cluster failed starting machines: machine \"7bdf6eb2-4c50-42f7-b3c2-c063caa69bad\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.64.198:22: i/o timeout\n"},{"name":"kubeadm.v1.35.1.calico.base","result":"FAIL","duration":1005795804775,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node: machine \"3bb47dc2-70f4-4c8a-8075-8a2a180c7a53\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 188.34.67.197:22: i/o timeout\n"}],"result":"FAIL","platform":"stackit","version":"4669.0.0"}
