{"tests":[{"name":"docker.base/docker-info","result":"PASS","duration":2316970700,"output":""},{"name":"docker.base/resources","result":"PASS","duration":3613116262,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.20MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:283bd538bf64595f345adcdaec227367cbe791edaddfad7265695c59a973c816 done\n        cluster.go:125: #5 naming to docker.io/library/sleep done\n        cluster.go:125: #5 DONE 0.0s\n        cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n        cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n"},{"name":"cl.cgroupv1/CgroupMounts","result":"PASS","duration":137226581,"output":""},{"name":"cl.cloudinit.multipart-mime","result":"PASS","duration":38534000820,"output":""},{"name":"cl.cgroupv1","result":"PASS","duration":39303912034,"output":"    --- PASS: cl.cgroupv1/CgroupMounts (0.14s)\n"},{"name":"cl.ignition.v2.xfsroot","result":"PASS","duration":40239305258,"output":""},{"name":"coreos.selinux.enforce","result":"PASS","duration":74229803320,"output":""},{"name":"docker.lib-coreos-dockerd-compat/docker-info","result":"PASS","duration":16317832295,"output":""},{"name":"docker.lib-coreos-dockerd-compat/resources","result":"PASS","duration":83041374637,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.3s done\n        cluster.go:125: #1 DONE 0.8s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 0.1s\n        cluster.go:125: #2 transferring context: 2B 0.2s done\n        cluster.go:125: #2 DONE 0.8s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.20MB 1.2s done\n        cluster.go:125: #3 DONE 1.7s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.5s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 0.8s done\n        cluster.go:125: #5 writing image sha256:4ac6817584861d2a510d20be47f9bfdbc129f26376d151335a04734db4f8d569 0.1s done\n        cluster.go:125: #5 naming to docker.io/library/sleep 0.1s done\n        cluster.go:125: #5 DONE 1.3s\n        cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n        cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n"},{"name":"cl.ignition.oem.indirect","result":"PASS","duration":183557991473,"output":""},{"name":"docker.base/networks-reliably","result":"PASS","duration":206286692164,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 4.04MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:95228fd81613b41035b3592dc22617d7e8cddd183e780bdfbcb007efb733b964 done\n        cluster.go:125: #5 naming to docker.io/library/ping done\n        cluster.go:125: #5 DONE 0.0s\n"},{"name":"docker.base/user-no-caps","result":"PASS","duration":5014335537,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B 0.0s done\n        cluster.go:125: #2 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 5.06MB 0.3s done\n        cluster.go:125: #3 DONE 0.3s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 0.2s done\n        cluster.go:125: #5 writing image sha256:774066f06ac497c5c5319ec345a8f1c1a6f01d9bd9d0d5588b714df2476ed65e 0.0s done\n        cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n        cluster.go:125: #5 DONE 0.3s\n"},{"name":"docker.base/ownership","result":"PASS","duration":15015259790,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/nginx\n        cluster.go:125: 9d11882d4b58: Pulling fs layer\n        cluster.go:125: 6a7080264fc9: Pulling fs layer\n        cluster.go:125: f176d5d8a1c1: Pulling fs layer\n        cluster.go:125: 63ef37274310: Pulling fs layer\n        cluster.go:125: 4744f2344932: Pulling fs layer\n        cluster.go:125: 39a24e95a1e9: Pulling fs layer\n        cluster.go:125: 400f8ac367e0: Pulling fs layer\n        cluster.go:125: 63ef37274310: Waiting\n        cluster.go:125: 4744f2344932: Waiting\n        cluster.go:125: 39a24e95a1e9: Waiting\n        cluster.go:125: 400f8ac367e0: Waiting\n        cluster.go:125: f176d5d8a1c1: Verifying Checksum\n        cluster.go:125: f176d5d8a1c1: Download complete\n        cluster.go:125: 63ef37274310: Verifying Checksum\n        cluster.go:125: 63ef37274310: Download complete\n        cluster.go:125: 6a7080264fc9: Verifying Checksum\n        cluster.go:125: 6a7080264fc9: Download complete\n        cluster.go:125: 4744f2344932: Verifying Checksum\n        cluster.go:125: 4744f2344932: Download complete\n        cluster.go:125: 39a24e95a1e9: Verifying Checksum\n        cluster.go:125: 39a24e95a1e9: Download complete\n        cluster.go:125: 9d11882d4b58: Verifying Checksum\n        cluster.go:125: 9d11882d4b58: Download complete\n        cluster.go:125: 400f8ac367e0: Verifying Checksum\n        cluster.go:125: 400f8ac367e0: Download complete\n        cluster.go:125: 9d11882d4b58: Pull complete\n        cluster.go:125: 6a7080264fc9: Pull complete\n        cluster.go:125: f176d5d8a1c1: Pull complete\n        cluster.go:125: 63ef37274310: Pull complete\n        cluster.go:125: 4744f2344932: Pull complete\n        cluster.go:125: 39a24e95a1e9: Pull complete\n        cluster.go:125: 400f8ac367e0: Pull complete\n        cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"docker.base","result":"PASS","duration":263112954902,"output":"    --- PASS: docker.base/docker-info (2.32s)\n    --- PASS: docker.base/resources (3.61s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 2.20MB 0.0s done\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.0s done\n            cluster.go:125: #5 writing image sha256:283bd538bf64595f345adcdaec227367cbe791edaddfad7265695c59a973c816 done\n            cluster.go:125: #5 naming to docker.io/library/sleep done\n            cluster.go:125: #5 DONE 0.0s\n            cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n            cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n    --- PASS: docker.base/networks-reliably (206.29s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 4.04MB 0.0s done\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.0s done\n            cluster.go:125: #5 writing image sha256:95228fd81613b41035b3592dc22617d7e8cddd183e780bdfbcb007efb733b964 done\n            cluster.go:125: #5 naming to docker.io/library/ping done\n            cluster.go:125: #5 DONE 0.0s\n    --- PASS: docker.base/user-no-caps (5.01s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B 0.0s done\n            cluster.go:125: #2 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 5.06MB 0.3s done\n            cluster.go:125: #3 DONE 0.3s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers\n            cluster.go:125: #5 exporting layers 0.2s done\n            cluster.go:125: #5 writing image sha256:774066f06ac497c5c5319ec345a8f1c1a6f01d9bd9d0d5588b714df2476ed65e 0.0s done\n            cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n            cluster.go:125: #5 DONE 0.3s\n    --- PASS: docker.base/ownership (15.02s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/nginx\n            cluster.go:125: 9d11882d4b58: Pulling fs layer\n            cluster.go:125: 6a7080264fc9: Pulling fs layer\n            cluster.go:125: f176d5d8a1c1: Pulling fs layer\n            cluster.go:125: 63ef37274310: Pulling fs layer\n            cluster.go:125: 4744f2344932: Pulling fs layer\n            cluster.go:125: 39a24e95a1e9: Pulling fs layer\n            cluster.go:125: 400f8ac367e0: Pulling fs layer\n            cluster.go:125: 63ef37274310: Waiting\n            cluster.go:125: 4744f2344932: Waiting\n            cluster.go:125: 39a24e95a1e9: Waiting\n            cluster.go:125: 400f8ac367e0: Waiting\n            cluster.go:125: f176d5d8a1c1: Verifying Checksum\n            cluster.go:125: f176d5d8a1c1: Download complete\n            cluster.go:125: 63ef37274310: Verifying Checksum\n            cluster.go:125: 63ef37274310: Download complete\n            cluster.go:125: 6a7080264fc9: Verifying Checksum\n            cluster.go:125: 6a7080264fc9: Download complete\n            cluster.go:125: 4744f2344932: Verifying Checksum\n            cluster.go:125: 4744f2344932: Download complete\n            cluster.go:125: 39a24e95a1e9: Verifying Checksum\n            cluster.go:125: 39a24e95a1e9: Download complete\n            cluster.go:125: 9d11882d4b58: Verifying Checksum\n            cluster.go:125: 9d11882d4b58: Download complete\n            cluster.go:125: 400f8ac367e0: Verifying Checksum\n            cluster.go:125: 400f8ac367e0: Download complete\n            cluster.go:125: 9d11882d4b58: Pull complete\n            cluster.go:125: 6a7080264fc9: Pull complete\n            cluster.go:125: f176d5d8a1c1: Pull complete\n            cluster.go:125: 63ef37274310: Pull complete\n            cluster.go:125: 4744f2344932: Pull complete\n            cluster.go:125: 39a24e95a1e9: Pull complete\n            cluster.go:125: 400f8ac367e0: Pull complete\n            cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"cl.ignition.translation","result":"PASS","duration":27751754452,"output":"        cluster.go:152: + ip --json address show kola | jq -r '.[] | .addr_info | .[] | select( .family == \"inet\") | .local'\n        cluster.go:152: + cat /etc/systemd/network/00-dummy.network\n"},{"name":"cl.ignition.v1.groups","result":"PASS","duration":296572804855,"output":""},{"name":"cl.locksmith.cluster","result":"PASS","duration":595905735906,"output":""},{"name":"systemd.journal.remote","result":"PASS","duration":491137730436,"output":"        cluster.go:125: Running as unit: systemd-journal-remote-client.service; invocation ID: 208e6e9b62de42a88ca539f044441fd9\n"},{"name":"coreos.locksmith.reboot","result":"PASS","duration":397206516577,"output":""},{"name":"cl.network.iptables","result":"PASS","duration":39281247701,"output":"        cluster.go:152: + sudo nft --json list ruleset | jq '.nftables[] | select(.rule) | .rule.expr[0].match.right'\n"},{"name":"cl.cloudinit.basic","result":"PASS","duration":43819024024,"output":""},{"name":"kubeadm.v1.33.8.flannel.cgroupv1.base","result":"FAIL","duration":1190045915416,"output":"        cluster.go:125: I0413 23:36:26.338609    1960 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0413 23:38:37.035449    2398 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.10\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 23:38:39.621394    2398 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.12]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 2.580717714s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.12:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 17.026482687s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 41.336469502s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 2m5.617805378s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 7b0u4r.02v4k7e8a3a1ortp\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.12:6443 --token 7b0u4r.02v4k7e8a3a1ortp \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:394d262b65b8902b347fc83e6bc2ade65b6a60fbd6999c5496bc43f5d681daeb \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        kubeadm.go:197: unable to setup cluster: unable to create worker node with large disk: machine \"9cc1895a-333f-42fd-bfca-335296fec439\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.23:22: connect: connection refused\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 9cc1895a-333f-42fd-bfca-335296fec439 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 9cc1895a-333f-42fd-bfca-335296fec439 console\n"},{"name":"docker.containerd-restart","result":"PASS","duration":174189961675,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Download complete\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n"},{"name":"cl.tpm.root","result":"PASS","duration":119814686971,"output":""},{"name":"cl.ignition.partition_on_boot_disk","result":"PASS","duration":45446077902,"output":"        filesystem.go:487: lsblk -f:\nNAME    FSTYPE   FSVER LABEL      UUID                                 FSAVAIL FSUSE% MOUNTPOINTS\nloop2   squashfs 4.0                                                                  \nloop3   squashfs 4.0                                                                  \nsr0                                                                                   \nvda                                                                                   \n|-vda1  vfat     FAT32 EFI-SYSTEM 0C46-7047                                           \n|-vda2                                                                                \n|-vda3  btrfs                     de1edd48-4571-4695-92f0-7af6e33c4e3d                \n| `-usr btrfs                     de1edd48-4571-4695-92f0-7af6e33c4e3d                /usr\n|-vda4                                                                                \n|-vda6  btrfs          OEM        7dd1319a-da93-42af-ac3b-f04d4587a8af    123M     1% /oem\n|-vda7                                                                                \n|-vda9  ext4     1.0   ROOT       e02793bf-3e0d-4c7e-b11a-92c664da7ce3    5.9G     0% /\n`-vda10 xfs            VAR        81ee1098-16f3-48a2-9b2c-2a45dd2cb636    5.2G     3% /var\n        filesystem.go:489: findmnt:\nTARGET                        SOURCE     FSTYPE     OPTIONS\n/                             /dev/vda9  ext4       rw,relatime,seclabel\n|-/boot                       systemd-1  autofs     rw,relatime,fd=33,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=7250\n|-/media                      tmpfs      tmpfs      rw,nosuid,nodev,noexec,relatime,seclabel\n|-/tmp                        tmpfs      tmpfs      rw,nosuid,nodev,seclabel,nr_inodes=1048576\n|-/dev                        devtmpfs   devtmpfs   rw,nosuid,seclabel,size=4096k,nr_inodes=303641,mode=755\n| |-/dev/hugepages            hugetlbfs  hugetlbfs  rw,nosuid,nodev,relatime,seclabel,pagesize=2M\n| |-/dev/mqueue               mqueue     mqueue     rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/dev/shm                  tmpfs      tmpfs      rw,nosuid,nodev,seclabel\n| `-/dev/pts                  devpts     devpts     rw,nosuid,noexec,relatime,seclabel,gid=5,mode=620,ptmxmode=000\n|-/sys                        sysfs      sysfs      rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/fs/selinux           selinuxfs  selinuxfs  rw,nosuid,noexec,relatime\n| |-/sys/kernel/debug         debugfs    debugfs    rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/kernel/tracing       tracefs    tracefs    rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/kernel/security      securityfs securityfs rw,nosuid,nodev,noexec,relatime\n| |-/sys/fs/cgroup            cgroup2    cgroup2    rw,nosuid,nodev,noexec,relatime,seclabel,nsdelegate,memory_recursiveprot\n| |-/sys/fs/pstore            pstore     pstore     rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/firmware/efi/efivars efivarfs   efivarfs   rw,nosuid,nodev,noexec,relatime\n| |-/sys/fs/bpf               bpf        bpf        rw,nosuid,nodev,noexec,relatime,mode=700\n| |-/sys/fs/fuse/connections  fusectl    fusectl    rw,nosuid,nodev,noexec,relatime\n| `-/sys/kernel/config        configfs   configfs   rw,nosuid,nodev,noexec,relatime\n|-/proc                       proc       proc       rw,nosuid,nodev,noexec,relatime\n| `-/proc/sys/fs/binfmt_misc  systemd-1  autofs     rw,relatime,fd=34,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=7254\n|-/run                        tmpfs      tmpfs      rw,nosuid,nodev,seclabel,size=494772k,nr_inodes=819200,mode=755\n| `-/run/user/500             tmpfs      tmpfs      rw,nosuid,nodev,relatime,seclabel,size=247384k,nr_inodes=61846,mode=700,uid=500,gid=500\n|-/var                        /dev/vda10 xfs        rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota\n|-/usr                        /dev/dm-0  btrfs      ro,relatime,seclabel,rescue=nologreplay,space_cache=v2,subvolid=5,subvol=/\n| `-/usr                      sysext     overlay    ro,nodev,relatime,seclabel,lowerdir=/run/systemd/sysext/meta/usr:/run/systemd/sysext/extensions/docker-flatcar/usr:/run/systemd/sysext/extensions/containerd-flatcar/usr:/usr\n|-/etc                        overlay    overlay    rw,noatime,seclabel,lowerdir=/sysroot/usr/share/flatcar/etc,upperdir=/sysroot/etc,workdir=/sysroot/.etc-work,uuid=on,metacopy=off\n`-/oem                        /dev/vda6  btrfs      rw,nodev,relatime,seclabel,discard=async,space_cache=v2,subvolid=5,subvol=/\n        cluster.go:152: + findmnt /var\n        cluster.go:152: + cat /var/hello\n"},{"name":"cl.ignition.luks","result":"PASS","duration":53046904481,"output":""},{"name":"kubeadm.v1.33.8.calico.base","result":"FAIL","duration":1272329358204,"output":"        cluster.go:125: I0413 23:45:29.803958    1820 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0413 23:47:39.346158    2150 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.10\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 23:47:49.960846    2150 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.25]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 17.640247923s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.25:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 23.444618236s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 40.678120647s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 4m38.24528918s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 5ed6qp.meai1twvgf9ln268\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.25:6443 --token 5ed6qp.meai1twvgf9ln268 \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:913092162eb08ea2bd560ee25a3492b023940022f00fd38e70d910fa65eacd31 \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: error: timed out waiting for the condition\n        kubeadm.go:197: unable to setup cluster: unable to run master script: Process exited with status 1\n"},{"name":"linux.nfs.v4","result":"PASS","duration":87868810733,"output":"        nfs.go:80: NFS server booted.\n        nfs.go:85: Test file \"/tmp/tmp.lHR05Qony2\" created on server.\n        nfs.go:125: NFS client booted.\n        nfs.go:133: Got NFS mount.\n"},{"name":"kubeadm.v1.33.8.calico.cgroupv1.base","result":"FAIL","duration":1507116069268,"output":"        cluster.go:125: I0413 23:58:25.273873    1976 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0413 23:59:37.880235    2318 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.10\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 23:59:41.451880    2318 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.40]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 9.588174687s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.40:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 12.351739674s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 34.955831959s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 2m13.717128573s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: i8f2km.0wnzweq9eaymvscf\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.40:6443 --token i8f2km.0wnzweq9eaymvscf \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:0b470487de2b45b5aea3bd402ae0b8157273a39e0d9384c248d0405a37182f8b \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: error: timed out waiting for the condition\n        kubeadm.go:197: unable to setup cluster: unable to run master script: Process exited with status 1\n"},{"name":"cl.overlay.cleanup","result":"PASS","duration":62859819481,"output":""},{"name":"kubeadm.v1.33.8.cilium.base/node_readiness","result":"FAIL","duration":202947462702,"output":"        kubeadm.go:213: nodes are not ready: ready nodes should be equal to 2: 1\n"},{"name":"cl.tpm.root-cryptenroll","result":"PASS","duration":79465446289,"output":""},{"name":"cl.ignition.v1.users","result":"PASS","duration":34353624426,"output":""},{"name":"kubeadm.v1.33.8.cilium.cgroupv1.base","result":"FAIL","duration":2064736922191,"output":"        cluster.go:125: I0413 23:36:31.213556    2025 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0413 23:38:36.387594    2443 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.10\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 23:38:39.325464    2443 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.10]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 6.132749432s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.10:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 21.592658625s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 56.257846675s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 6m21.396181396s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: gcl620.b2dy8i88x66s839t\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.10:6443 --token gcl620.b2dy8i88x66s839t \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:d3b2f50711124bc1106750c4df97f4bfed65978f4eb8e1566098f5b589fb74e9 \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0414 00:02:07.106531    1879 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: error execution phase preflight: couldn't validate the identity of the API Server: could not find a JWS signature in the cluster-info ConfigMap for token ID \"9ddgh7\"\n        cluster.go:125: To see the stack trace of this error execute with --v=5 or higher\n        kubeadm.go:197: unable to setup cluster: unable to run worker script: Process exited with status 1\n"},{"name":"cl.network.wireguard","result":"PASS","duration":29808454052,"output":"        cluster.go:152: + ip --json address show kv0 | jq -r '.[] | .addr_info | .[] | select( .family == \"inet\") | .local'\n"},{"name":"cl.sysext.fallbackdownload","result":"SKIP","duration":28991780240,"output":"        update.go:652: Test skipped, only dev builds are supported\n"},{"name":"kubeadm.v1.33.8.cilium.base/nginx_deployment","result":"FAIL","duration":197971329606,"output":"        kubeadm.go:232: nginx is not deployed: ready replicas should be equal to 1: null\n"},{"name":"docker.selinux","result":"PASS","duration":35643775418,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Download complete\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: sh: can't create /opt/hello: Permission denied\n"},{"name":"coreos.tls.fetch-urls","result":"PASS","duration":33331360891,"output":""},{"name":"kubeadm.v1.35.1.flannel.base/node_readiness","result":"PASS","duration":49324714431,"output":""},{"name":"kubeadm.v1.35.1.flannel.base/nginx_deployment","result":"PASS","duration":52639488347,"output":""},{"name":"cl.ignition.v1.once","result":"PASS","duration":57765373080,"output":""},{"name":"kubeadm.v1.33.8.cilium.base/NFS_deployment","result":"PASS","duration":108395064772,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.33.8.cilium.base/IPSec_encryption","result":"FAIL","duration":69307313579,"output":"        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: Error: Unable to determine status:  timeout while waiting for status to become successful: context deadline exceeded\n        cluster.go:145: \"/opt/bin/cilium status --wait --wait-duration 1m\" failed: output \u001b[33m    /¯¯\\\n\u001b[36m /¯¯\u001b[33m\\__/\u001b[32m¯¯\\\u001b[0m    Cilium:         \u001b[31m2 errors\u001b[0m\n\u001b[36m \\__\u001b[31m/¯¯\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n\u001b[32m /¯¯\u001b[31m\\__/\u001b[35m¯¯\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n\u001b[32m \\__\u001b[34m/¯¯\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n\u001b[34m    \\__/\n\u001b[0m\nDeployment        cilium-operator    Desired: 1, Ready: \u001b[32m1/1\u001b[0m, Available: \u001b[32m1/1\u001b[0m\nDaemonSet         cilium             Desired: 2, Ready: \u001b[33m1/2\u001b[0m, Available: \u001b[33m1/2\u001b[0m, Unavailable: \u001b[31m1/2\u001b[0m\nContainers:       cilium             Running: \u001b[32m2\u001b[0m\n                  cilium-operator    Running: \u001b[32m1\u001b[0m\nCluster Pods:     5/5 managed by Cilium\nImage versions    cilium-operator    quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: 1\n                  cilium             quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: 2\nErrors:           cilium             cilium          1 pods of DaemonSet cilium are not ready\n                  cilium             cilium-q4p6f    unable to retrieve cilium status: command terminated with exit code 1, status Process exited with status 1\n"},{"name":"kubeadm.v1.33.8.cilium.base","result":"FAIL","duration":1009904912114,"output":"        cluster.go:125: I0413 23:56:55.230342    1828 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0413 23:57:17.020754    2066 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.10\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 23:57:17.327386    2066 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.37]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 2.009644974s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.37:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 4.620009323s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 5.528001115s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 10.009178186s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: t25qvd.cw1m5whx3hkm9b6p\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.37:6443 --token t25qvd.cw1m5whx3hkm9b6p \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:de56df9c1fd4519727d0da19a8ea27edee0a6e892d371304fee9dd0644e42063 \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0414 00:00:28.091535    1689 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- FAIL: kubeadm.v1.33.8.cilium.base/node_readiness (202.95s)\n            kubeadm.go:213: nodes are not ready: ready nodes should be equal to 2: 1\n    --- FAIL: kubeadm.v1.33.8.cilium.base/nginx_deployment (197.97s)\n            kubeadm.go:232: nginx is not deployed: ready replicas should be equal to 1: null\n    --- PASS: kubeadm.v1.33.8.cilium.base/NFS_deployment (108.40s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n    --- FAIL: kubeadm.v1.33.8.cilium.base/IPSec_encryption (69.31s)\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: Error: Unable to determine status:  timeout while waiting for status to become successful: context deadline exceeded\n            cluster.go:145: \"/opt/bin/cilium status --wait --wait-duration 1m\" failed: output \u001b[33m    /¯¯\\\n    \u001b[36m /¯¯\u001b[33m\\__/\u001b[32m¯¯\\\u001b[0m    Cilium:         \u001b[31m2 errors\u001b[0m\n    \u001b[36m \\__\u001b[31m/¯¯\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n    \u001b[32m /¯¯\u001b[31m\\__/\u001b[35m¯¯\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n    \u001b[32m \\__\u001b[34m/¯¯\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n    \u001b[34m    \\__/\n    \u001b[0m\n    Deployment        cilium-operator    Desired: 1, Ready: \u001b[32m1/1\u001b[0m, Available: \u001b[32m1/1\u001b[0m\n    DaemonSet         cilium             Desired: 2, Ready: \u001b[33m1/2\u001b[0m, Available: \u001b[33m1/2\u001b[0m, Unavailable: \u001b[31m1/2\u001b[0m\n    Containers:       cilium             Running: \u001b[32m2\u001b[0m\n                      cilium-operator    Running: \u001b[32m1\u001b[0m\n    Cluster Pods:     5/5 managed by Cilium\n    Image versions    cilium-operator    quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: 1\n                      cilium             quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: 2\n    Errors:           cilium             cilium          1 pods of DaemonSet cilium are not ready\n                      cilium             cilium-q4p6f    unable to retrieve cilium status: command terminated with exit code 1, status Process exited with status 1\n"},{"name":"cl.internet/UpdateEngine","result":"PASS","duration":308504814,"output":""},{"name":"kubeadm.v1.35.1.flannel.base/NFS_deployment","result":"PASS","duration":113550792169,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"cl.internet/DockerPing","result":"PASS","duration":7817778011,"output":""},{"name":"cl.internet/DockerEcho","result":"PASS","duration":809784575,"output":""},{"name":"kubeadm.v1.35.1.flannel.base","result":"PASS","duration":965280360245,"output":"        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.13.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.6-0\n        cluster.go:125: [init] Using Kubernetes version: v1.35.3\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 23:58:39.326683    2035 checks.go:906] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.39]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.526457994s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.39:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 6.04799141s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 25.687377289s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 1m7.959057914s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 8nmrtj.xq3pdy24ocxmxzgt\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.39:6443 --token 8nmrtj.xq3pdy24ocxmxzgt \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:39a18086912b6be177cff9436420c291a993fbb69d898960c13a5609acfa02e9 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0414 00:08:40.503264    1635 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.35.1.flannel.base/node_readiness (49.32s)\n    --- PASS: kubeadm.v1.35.1.flannel.base/nginx_deployment (52.64s)\n    --- PASS: kubeadm.v1.35.1.flannel.base/NFS_deployment (113.55s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"docker.lib-coreos-dockerd-compat/networks-reliably","result":"PASS","duration":2169515339585,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.7s done\n        cluster.go:125: #1 DONE 1.2s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B 0.3s done\n        cluster.go:125: #2 DONE 0.7s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 4.04MB 2.6s done\n        cluster.go:125: #3 DONE 3.4s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.8s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 1.3s done\n        cluster.go:125: #5 writing image sha256:96e8da1bd03865c5f5d6445e0eac330bf31d20083a63fd074dcf9c0db0df209a 0.0s done\n        cluster.go:125: #5 naming to docker.io/library/ping 0.1s done\n        cluster.go:125: #5 DONE 1.6s\n"},{"name":"cl.internet/NTPDate","result":"PASS","duration":6981143811,"output":""},{"name":"docker.lib-coreos-dockerd-compat/user-no-caps","result":"PASS","duration":3283413469,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 5.06MB 0.1s done\n        cluster.go:125: #3 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:fb975546e9d5bf57e07f3edbe9afe9bb0f9478e0d5ca642b283807c3ff0dd922 done\n        cluster.go:125: #5 naming to docker.io/library/captest done\n        cluster.go:125: #5 DONE 0.1s\n"},{"name":"cl.internet","result":"PASS","duration":42076312746,"output":"    --- PASS: cl.internet/UpdateEngine (0.31s)\n    --- PASS: cl.internet/DockerPing (7.82s)\n    --- PASS: cl.internet/DockerEcho (0.81s)\n    --- PASS: cl.internet/NTPDate (6.98s)\n"},{"name":"cl.sysext.boot","result":"PASS","duration":126905611501,"output":"        cluster.go:125: + sudo systemctl mask --now ensure-sysext\n        cluster.go:125: Created symlink /etc/systemd/system/ensure-sysext.service → /dev/null.\n        cluster.go:125: + sudo systemctl mask --now systemd-sysext\n        cluster.go:125: Created symlink /etc/systemd/system/systemd-sysext.service → /dev/null.\n        cluster.go:125: + sudo mkdir -p /etc/flatcar/sysext /etc/flatcar/oem-sysext /oem/sysext /etc/extensions\n        cluster.go:125: + sudo tee /oem/oem-release\n        cluster.go:125: + echo ID=test\n        cluster.go:125: + echo myext\n        cluster.go:125: + sudo tee /etc/flatcar/enabled-sysext.conf\n        cluster.go:125: + sudo touch /oem/sysext/active-oem-test /etc/flatcar/oem-sysext/oem-test-4081.3.7.raw /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw /etc/flatcar/sysext/flatcar-myext-1.2.3.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /etc/extensions/oem-test.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/sysext/flatcar-myext-1.2.3.raw /etc/extensions/flatcar-myext.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4081.3.7.raw = /oem/sysext/oem-test-4081.3.7.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw = /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw ']'\n        cluster.go:125: + sudo mv /oem/sysext/oem-test-4081.3.7.raw /etc/flatcar/oem-sysext/\n        cluster.go:125: + sudo mv /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /oem/sysext/\n        cluster.go:125: + sudo ln -fs /oem/sysext/oem-test-1.2.3.raw /etc/extensions/oem-test.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/sysext/flatcar-myext-1.2.3.raw /etc/extensions/flatcar-myext.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4081.3.7.raw = /oem/sysext/oem-test-4081.3.7.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw = /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw ']'\n        cluster.go:125: + '[' -e /etc/flatcar/oem-sysext/oem-test-1.2.3.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4081.3.7.raw = /oem/sysext/oem-test-4081.3.7.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw = /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw ']'\n"},{"name":"cl.ignition.symlink","result":"PASS","duration":29443463029,"output":"        cluster.go:152: + readlink /etc/localtime\n"},{"name":"docker.lib-coreos-dockerd-compat/ownership","result":"PASS","duration":43115040378,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/nginx\n        cluster.go:125: 9d11882d4b58: Pulling fs layer\n        cluster.go:125: 6a7080264fc9: Pulling fs layer\n        cluster.go:125: f176d5d8a1c1: Pulling fs layer\n        cluster.go:125: 63ef37274310: Pulling fs layer\n        cluster.go:125: 4744f2344932: Pulling fs layer\n        cluster.go:125: 39a24e95a1e9: Pulling fs layer\n        cluster.go:125: 400f8ac367e0: Pulling fs layer\n        cluster.go:125: 4744f2344932: Waiting\n        cluster.go:125: 39a24e95a1e9: Waiting\n        cluster.go:125: 400f8ac367e0: Waiting\n        cluster.go:125: 63ef37274310: Waiting\n        cluster.go:125: f176d5d8a1c1: Download complete\n        cluster.go:125: 63ef37274310: Verifying Checksum\n        cluster.go:125: 63ef37274310: Download complete\n        cluster.go:125: 9d11882d4b58: Verifying Checksum\n        cluster.go:125: 9d11882d4b58: Download complete\n        cluster.go:125: 6a7080264fc9: Verifying Checksum\n        cluster.go:125: 6a7080264fc9: Download complete\n        cluster.go:125: 39a24e95a1e9: Verifying Checksum\n        cluster.go:125: 39a24e95a1e9: Download complete\n        cluster.go:125: 4744f2344932: Download complete\n        cluster.go:125: 400f8ac367e0: Verifying Checksum\n        cluster.go:125: 400f8ac367e0: Download complete\n        cluster.go:125: 9d11882d4b58: Pull complete\n        cluster.go:125: 6a7080264fc9: Pull complete\n        cluster.go:125: f176d5d8a1c1: Pull complete\n        cluster.go:125: 63ef37274310: Pull complete\n        cluster.go:125: 4744f2344932: Pull complete\n        cluster.go:125: 39a24e95a1e9: Pull complete\n        cluster.go:125: 400f8ac367e0: Pull complete\n        cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"docker.network-openbsd-nc","result":"PASS","duration":43300700660,"output":"        docker.go:413: creating netcat containers\n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 78B 0.0s\n        cluster.go:125: #1 transferring dockerfile: 108B 0.1s done\n        cluster.go:125: #1 DONE 0.3s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 0.0s\n        cluster.go:125: #2 transferring context: 2B 0.0s done\n        cluster.go:125: #2 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.45MB 0.4s done\n        cluster.go:125: #3 DONE 0.6s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:9d7c2a6d23efd4fa5b692a8647d3e6b56332ab74dca480a451375f86ce598149 0.0s done\n        cluster.go:125: #5 naming to docker.io/library/netcat 0.0s done\n        cluster.go:125: #5 DONE 0.2s\n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 78B 0.0s\n        cluster.go:125: #1 transferring dockerfile: 108B 0.1s done\n        cluster.go:125: #1 DONE 0.3s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B 0.0s done\n        cluster.go:125: #2 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.45MB 0.1s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:05725aa047263bfabc8045a9df2c900207d5444c0fc647d3e813126e65227166 done\n        cluster.go:125: #5 naming to docker.io/library/netcat 0.1s done\n        cluster.go:125: #5 DONE 0.2s\n"},{"name":"docker.lib-coreos-dockerd-compat","result":"PASS","duration":2368284873927,"output":"    --- PASS: docker.lib-coreos-dockerd-compat/docker-info (16.32s)\n    --- PASS: docker.lib-coreos-dockerd-compat/resources (83.04s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B 0.3s done\n            cluster.go:125: #1 DONE 0.8s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 0.1s\n            cluster.go:125: #2 transferring context: 2B 0.2s done\n            cluster.go:125: #2 DONE 0.8s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 2.20MB 1.2s done\n            cluster.go:125: #3 DONE 1.7s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.5s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers\n            cluster.go:125: #5 exporting layers 0.8s done\n            cluster.go:125: #5 writing image sha256:4ac6817584861d2a510d20be47f9bfdbc129f26376d151335a04734db4f8d569 0.1s done\n            cluster.go:125: #5 naming to docker.io/library/sleep 0.1s done\n            cluster.go:125: #5 DONE 1.3s\n            cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n            cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n    --- PASS: docker.lib-coreos-dockerd-compat/networks-reliably (2169.52s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B 0.7s done\n            cluster.go:125: #1 DONE 1.2s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context:\n            cluster.go:125: #2 transferring context: 2B 0.3s done\n            cluster.go:125: #2 DONE 0.7s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 4.04MB 2.6s done\n            cluster.go:125: #3 DONE 3.4s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.8s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers\n            cluster.go:125: #5 exporting layers 1.3s done\n            cluster.go:125: #5 writing image sha256:96e8da1bd03865c5f5d6445e0eac330bf31d20083a63fd074dcf9c0db0df209a 0.0s done\n            cluster.go:125: #5 naming to docker.io/library/ping 0.1s done\n            cluster.go:125: #5 DONE 1.6s\n    --- PASS: docker.lib-coreos-dockerd-compat/user-no-caps (3.28s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 5.06MB 0.1s done\n            cluster.go:125: #3 DONE 0.2s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.1s done\n            cluster.go:125: #5 writing image sha256:fb975546e9d5bf57e07f3edbe9afe9bb0f9478e0d5ca642b283807c3ff0dd922 done\n            cluster.go:125: #5 naming to docker.io/library/captest done\n            cluster.go:125: #5 DONE 0.1s\n    --- PASS: docker.lib-coreos-dockerd-compat/ownership (43.12s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/nginx\n            cluster.go:125: 9d11882d4b58: Pulling fs layer\n            cluster.go:125: 6a7080264fc9: Pulling fs layer\n            cluster.go:125: f176d5d8a1c1: Pulling fs layer\n            cluster.go:125: 63ef37274310: Pulling fs layer\n            cluster.go:125: 4744f2344932: Pulling fs layer\n            cluster.go:125: 39a24e95a1e9: Pulling fs layer\n            cluster.go:125: 400f8ac367e0: Pulling fs layer\n            cluster.go:125: 4744f2344932: Waiting\n            cluster.go:125: 39a24e95a1e9: Waiting\n            cluster.go:125: 400f8ac367e0: Waiting\n            cluster.go:125: 63ef37274310: Waiting\n            cluster.go:125: f176d5d8a1c1: Download complete\n            cluster.go:125: 63ef37274310: Verifying Checksum\n            cluster.go:125: 63ef37274310: Download complete\n            cluster.go:125: 9d11882d4b58: Verifying Checksum\n            cluster.go:125: 9d11882d4b58: Download complete\n            cluster.go:125: 6a7080264fc9: Verifying Checksum\n            cluster.go:125: 6a7080264fc9: Download complete\n            cluster.go:125: 39a24e95a1e9: Verifying Checksum\n            cluster.go:125: 39a24e95a1e9: Download complete\n            cluster.go:125: 4744f2344932: Download complete\n            cluster.go:125: 400f8ac367e0: Verifying Checksum\n            cluster.go:125: 400f8ac367e0: Download complete\n            cluster.go:125: 9d11882d4b58: Pull complete\n            cluster.go:125: 6a7080264fc9: Pull complete\n            cluster.go:125: f176d5d8a1c1: Pull complete\n            cluster.go:125: 63ef37274310: Pull complete\n            cluster.go:125: 4744f2344932: Pull complete\n            cluster.go:125: 39a24e95a1e9: Pull complete\n            cluster.go:125: 400f8ac367e0: Pull complete\n            cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"cl.update.reboot","result":"PASS","duration":57948572598,"output":""},{"name":"systemd.journal.user","result":"PASS","duration":27423591492,"output":""},{"name":"coreos.update.badusr","result":"PASS","duration":383805108772,"output":""},{"name":"bpf.ig/ig","result":"PASS","duration":59924969445,"output":"        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n        cluster.go:125: time=\"2026-04-14T00:14:03Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Get \\\"http://%2Frun%2Fdocker.sock/v1.51/containers/json?all=1\\\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\"\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + docker info\n        cluster.go:125: + docker ps\n        cluster.go:125: + docker images\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + dig kinvolk.io\n        cluster.go:125: + dig flatcar.org\n        cluster.go:125: + dig stable.release.flatcar-linux.net\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"bpf.ig","result":"PASS","duration":62094785182,"output":"    --- PASS: bpf.ig/ig (59.92s)\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n            cluster.go:125: time=\"2026-04-14T00:14:03Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Get \\\"http://%2Frun%2Fdocker.sock/v1.51/containers/json?all=1\\\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\"\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + docker info\n            cluster.go:125: + docker ps\n            cluster.go:125: + docker images\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + dig kinvolk.io\n            cluster.go:125: + dig flatcar.org\n            cluster.go:125: + dig stable.release.flatcar-linux.net\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"coreos.ignition.resource.s3.versioned","result":"PASS","duration":30120319387,"output":""},{"name":"cl.ignition.instantiated.enable-unit","result":"PASS","duration":32594636029,"output":""},{"name":"cl.flannel.vxlan","result":"PASS","duration":205576890067,"output":"        flannel.go:121: ping from 536a4292-a13a-4331-97de-c5aa5577c73b(10.254.21.0) to c1497dc7-6a1f-43bf-a578-719ba5f3e012(10.254.72.0)\n"},{"name":"cl.etcd-member.v2-backup-restore","result":"PASS","duration":88935129999,"output":"        cluster.go:125: 2026-04-14T00:16:47Z\tinfo\tetcdutl/backup_command.go:216\tignoring EntryConfChange raft entry\n        cluster.go:125: 2026-04-14T00:16:47Z\tinfo\tetcdutl/backup_command.go:231\tignoring member attribute update on\t{\"entry\": \"Term:2 Index:3 Data:\\\"\\\\010\\\\202\\\\210\\\\242\\\\306\\\\225\\\\261\\\\247\\\\313\\\\220\\\\001\\\\022\\\\003PUT\\\\032\u0026/0/members/88a5f94fb2299096/attributes\\\\\\\"R{\\\\\\\"name\\\\\\\":\\\\\\\"7df8edf4616642a3a66ef59a94f555b5\\\\\\\",\\\\\\\"clientURLs\\\\\\\":[\\\\\\\"http://10.0.0.76:2379\\\\\\\"]}(\\\\0002\\\\0008\\\\000H\\\\000P\\\\000X\\\\000`\\\\000h\\\\000p\\\\000x\\\\000\\\\200\\\\001\\\\000\\\" \", \"v2Req.Path\": \"/0/members/88a5f94fb2299096/attributes\"}\n        cluster.go:125: 2026-04-14T00:16:47Z\tinfo\tetcdutl/backup_command.go:252\tignoring v3 raft entry\n        cluster.go:125: 2026-04-14T00:16:47Z\tinfo\tmembership/store.go:119\tTrimming membership information from the backend...\n"},{"name":"cl.ignition.oem.regular","result":"PASS","duration":59754826302,"output":""},{"name":"packages/sys-cluster/ipvsadm","result":"PASS","duration":18458947337,"output":""},{"name":"devcontainer.docker","result":"FAIL","duration":2999707299858,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + ARCH=amd64\n        cluster.go:125: + VERSION=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + EXPECTED_VERSION=4081.3.7\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-17410\n        cluster.go:125: + USR_SRC_DIR=/home/core/dev-container-workdir-17410/src\n        cluster.go:125: + VAR_TMP_DIR=/home/core/dev-container-workdir-17410/tmp\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-17410/src /home/core/dev-container-workdir-17410/tmp\n        cluster.go:125: ++ parted flatcar_developer_container.bin unit b print\n        cluster.go:125: ++ grep Start --after-context=1\n        cluster.go:125: ++ tail --lines=1\n        cluster.go:125: ++ awk '{ print $2 }'\n        cluster.go:125: ++ head --bytes=-2\n        cluster.go:125: + offset=2097152\n        cluster.go:125: + mkdir root\n        cluster.go:125: + sudo mount -o loop,ro,offset=2097152 flatcar_developer_container.bin root\n        cluster.go:125: + sudo tar -C root -czf dev-container-image.tar.gz .\n        cluster.go:125: + sudo umount root\n        cluster.go:125: + rm -f flatcar_developer_container.bin\n        cluster.go:125: + docker import dev-container-image.tar.gz dev-container:42\n        cluster.go:125: + sudo rm -f dev-container-image.tar.gz\n        cluster.go:125: + restorecon /home/core/dev-container-script\n        cluster.go:125: + docker run --log-driver=journald --env PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs --env EXPECTED_VERSION=4081.3.7 --mount type=bind,source=/lib/modules,target=/lib/modules,readonly=true --mount type=bind,source=/home/core/dev-container-script,target=/home/core/dev-container-script,readonly=true --mount type=bind,source=/home/core/dev-container-workdir-17410/src,target=/usr/src --mount type=bind,source=/home/core/dev-container-workdir-17410/tmp,target=/var/tmp dev-container:42 /bin/bash /home/core/dev-container-script\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + [[ 4081.3.7 != \\4\\0\\8\\1\\.\\3\\.\\7 ]]\n        cluster.go:125: + export PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + export 'FEATURES=-ipc-sandbox -network-sandbox'\n        cluster.go:125: + FEATURES='-ipc-sandbox -network-sandbox'\n        cluster.go:125: + emerge-gitclone\n        cluster.go:125: Cloning into '/var/lib/portage/scripts'...\n        cluster.go:125: error: RPC failed; curl 56 Recv failure: Connection reset by peer\n        cluster.go:125: error: 1333 bytes of body are still expected\n        cluster.go:125: fetch-pack: unexpected disconnect while reading sideband packet\n        cluster.go:125: fatal: early EOF\n        cluster.go:125: fatal: fetch-pack: invalid index-pack output\n        cluster.go:125: \u003e\u003e\u003e No git repositories configured.\n        devcontainer.go:236: main script failed: Process exited with status 1\n"},{"name":"docker.userns","result":"PASS","duration":53403131495,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.24MB 0.0s done\n        cluster.go:125: #3 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:a9842ae92033a222102fe233cae1381c1c801158d2c967d647583bc606e3b65b\n        cluster.go:125: #5 writing image sha256:a9842ae92033a222102fe233cae1381c1c801158d2c967d647583bc606e3b65b done\n        cluster.go:125: #5 naming to docker.io/library/userns-test 0.1s done\n        cluster.go:125: #5 DONE 0.2s\n"},{"name":"kubeadm.v1.33.8.flannel.base/node_readiness","result":"PASS","duration":258191815680,"output":""},{"name":"cl.disk.raid0.data","result":"PASS","duration":85280052832,"output":""},{"name":"cl.basic/ReadOnly","result":"PASS","duration":726901531,"output":""},{"name":"cl.basic/Useradd","result":"PASS","duration":1078273413,"output":""},{"name":"cl.basic/MachineID","result":"PASS","duration":497593688,"output":""},{"name":"cl.basic/CloudConfig","result":"PASS","duration":1821069295,"output":""},{"name":"cl.basic/Script","result":"PASS","duration":1129252200,"output":""},{"name":"cl.basic/PortSSH","result":"PASS","duration":555830349,"output":""},{"name":"cl.basic/RandomUUID","result":"PASS","duration":770058883,"output":""},{"name":"cl.basic/Microcode","result":"PASS","duration":628003310,"output":""},{"name":"cl.basic/DbusPerms","result":"PASS","duration":829580172,"output":""},{"name":"cl.basic/Symlink","result":"PASS","duration":456523840,"output":""},{"name":"cl.basic/SymlinkFlatcar","result":"PASS","duration":574998642,"output":""},{"name":"cl.basic/UpdateEngineKeys","result":"PASS","duration":607981406,"output":""},{"name":"cl.basic/ServicesActive","result":"PASS","duration":821092806,"output":""},{"name":"cl.basic","result":"PASS","duration":43624869469,"output":"    --- PASS: cl.basic/ReadOnly (0.73s)\n    --- PASS: cl.basic/Useradd (1.08s)\n    --- PASS: cl.basic/MachineID (0.50s)\n    --- PASS: cl.basic/CloudConfig (1.82s)\n    --- PASS: cl.basic/Script (1.13s)\n    --- PASS: cl.basic/PortSSH (0.56s)\n    --- PASS: cl.basic/RandomUUID (0.77s)\n    --- PASS: cl.basic/Microcode (0.63s)\n    --- PASS: cl.basic/DbusPerms (0.83s)\n    --- PASS: cl.basic/Symlink (0.46s)\n    --- PASS: cl.basic/SymlinkFlatcar (0.57s)\n    --- PASS: cl.basic/UpdateEngineKeys (0.61s)\n    --- PASS: cl.basic/ServicesActive (0.82s)\n"},{"name":"kubeadm.v1.34.4.cilium.base","result":"FAIL","duration":789147820905,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node with large disk: machine \"3c967970-51f1-4992-8571-db1195805b76\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.77:22: connect: no route to host\n"},{"name":"kubeadm.v1.34.4.flannel.base","result":"FAIL","duration":763455021842,"output":"        cluster.go:125: I0414 00:15:31.366935    1794 version.go:260] remote version is much newer: v1.35.3; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.6\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.6\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.6\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.6\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0414 00:15:53.321714    2031 version.go:260] remote version is much newer: v1.35.3; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.6\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0414 00:15:53.723467    2031 checks.go:827] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.73]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 2.007438221s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.73:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 5.411500533s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 7.711099252s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 15.502996971s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: cnutst.b3i1tlyw3og9d8p4\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.73:6443 --token cnutst.b3i1tlyw3og9d8p4 \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:469fe83a6b695f822d8656bef06d047671c08b4cf3d48fb5b381a922e18e5c77 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        kubeadm.go:197: unable to setup cluster: unable to create worker node with large disk: machine \"e72c928e-d7f2-412d-885a-47fb4027a771\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.78:22: connect: no route to host\n"},{"name":"cl.ignition.v1.sethostname","result":"PASS","duration":41072934994,"output":""},{"name":"packages/sys-block/open-iscsi","result":"PASS","duration":459054365536,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/targetcli-fb:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/targetcli-fb\n        cluster.go:125: a2318d6c47ec: Pulling fs layer\n        cluster.go:125: 3d3086a1439f: Pulling fs layer\n        cluster.go:125: a2318d6c47ec: Verifying Checksum\n        cluster.go:125: a2318d6c47ec: Download complete\n        cluster.go:125: 3d3086a1439f: Verifying Checksum\n        cluster.go:125: 3d3086a1439f: Download complete\n        cluster.go:125: a2318d6c47ec: Pull complete\n        cluster.go:125: 3d3086a1439f: Pull complete\n        cluster.go:125: Digest: sha256:b6cd65db981974e8b74938617218dd023775b969f9a059ced21e6ce6fa4763c1\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/targetcli-fb:latest\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: Created symlink /etc/systemd/system/remote-fs.target.wants/iscsi.service → /usr/lib/systemd/system/iscsi.service.\n"},{"name":"packages","result":"PASS","duration":589689560599,"output":"    --- PASS: packages/sys-cluster/ipvsadm (18.46s)\n    --- PASS: packages/sys-block/open-iscsi (459.05s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/targetcli-fb:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/targetcli-fb\n            cluster.go:125: a2318d6c47ec: Pulling fs layer\n            cluster.go:125: 3d3086a1439f: Pulling fs layer\n            cluster.go:125: a2318d6c47ec: Verifying Checksum\n            cluster.go:125: a2318d6c47ec: Download complete\n            cluster.go:125: 3d3086a1439f: Verifying Checksum\n            cluster.go:125: 3d3086a1439f: Download complete\n            cluster.go:125: a2318d6c47ec: Pull complete\n            cluster.go:125: 3d3086a1439f: Pull complete\n            cluster.go:125: Digest: sha256:b6cd65db981974e8b74938617218dd023775b969f9a059ced21e6ce6fa4763c1\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/targetcli-fb:latest\n            cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n            cluster.go:125: Created symlink /etc/systemd/system/remote-fs.target.wants/iscsi.service → /usr/lib/systemd/system/iscsi.service.\n"},{"name":"kubeadm.v1.33.8.flannel.base/nginx_deployment","result":"PASS","duration":176131491353,"output":""},{"name":"systemd.sysusers.gshadow","result":"PASS","duration":53654997874,"output":""},{"name":"cl.tang.root","result":"PASS","duration":137827550462,"output":"        tang.go:148: Started tang on 10.0.0.1:45579\n"},{"name":"cl.etcd-member.discovery","result":"PASS","duration":137089975953,"output":""},{"name":"sysext.disable-containerd","result":"PASS","duration":42701365943,"output":""},{"name":"sysext.custom-oem","result":"PASS","duration":187323575198,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-19850\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-19850\n        cluster.go:125: + sudo systemd-nspawn --console=pipe --bind-ro=/home/core/dev-container-script.sh --bind=/home/core/dev-container-workdir-19850:/work --image=flatcar_developer_container.bin --machine=flatcar-developer-container /bin/bash /home/core/dev-container-script.sh\n        cluster.go:125: + set -euo pipefail\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-13-1903\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-13-1903\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version_id=4081.3.7\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64-usr\n        cluster.go:125: + board=amd64-usr\n        cluster.go:125: + mkdir -p /work/sysext_rootfs/usr/share/flatcar-sysext-kola-test\n        cluster.go:125: + echo 4081.3.7\n        cluster.go:125: + mkdir -p /work/sysext_rootfs/usr/lib/extension-release.d\n        cluster.go:125: + sysext_arch=x86-64\n        cluster.go:125: + [[ amd64-usr = \\a\\r\\m\\6\\4\\-\\u\\s\\r ]]\n        cluster.go:125: + metadata=('ID=flatcar' \"VERSION_ID=${version_id}\" \"ARCHITECTURE=${sysext_arch}\")\n        cluster.go:125: + metadata_file=/work/sysext_rootfs/usr/lib/extension-release.d/extension-release.oem-test\n        cluster.go:125: + printf '%s\\n' ID=flatcar VERSION_ID=4081.3.7 ARCHITECTURE=x86-64\n        cluster.go:125: + mksquashfs /work/sysext_rootfs /work/oem-test-4081.3.7.raw -all-root\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-13-1915\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: + sysext_file=/home/core/dev-container-workdir-19850/oem-test-4081.3.7.raw\n        cluster.go:125: + [[ ! -e /home/core/dev-container-workdir-19850/oem-test-4081.3.7.raw ]]\n        cluster.go:125: + [[ ! -e /oem/oem-release ]]\n        cluster.go:125: + printf '%s\\n' ID=test VERSION_ID=1.0.0 'NAME=testing stuff'\n        cluster.go:125: + sudo tee /oem/oem-release\n        cluster.go:125: + sudo mkdir -p /oem/sysext\n        cluster.go:125: + sudo mv /home/core/dev-container-workdir-19850/oem-test-4081.3.7.raw /oem/sysext\n        cluster.go:125: + sudo touch /oem/sysext/active-oem-test\n        cluster.go:125: + sudo flatcar-reset --keep-machine-id --keep-paths /var/log\n        cluster.go:125: + set -euo pipefail\n        cluster.go:125: ++ systemd-sysext list --json=pretty\n        cluster.go:125: + list_out='[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"containerd-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/containerd-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776108218939451\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"docker-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/docker-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776108247798319\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"oem-test\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/oem-test.raw\",\n        cluster.go:125: \t\t\"time\" : 1776126551832074\n        cluster.go:125: \t}\n        cluster.go:125: ]'\n        cluster.go:125: ++ systemd-sysext status --json=pretty\n        cluster.go:125: + status_out='[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/opt\",\n        cluster.go:125: \t\t\"extensions\" : \"none\",\n        cluster.go:125: \t\t\"since\" : null\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/usr\",\n        cluster.go:125: \t\t\"extensions\" : [\n        cluster.go:125: \t\t\t\"containerd-flatcar\",\n        cluster.go:125: \t\t\t\"docker-flatcar\",\n        cluster.go:125: \t\t\t\"oem-test\"\n        cluster.go:125: \t\t],\n        cluster.go:125: \t\t\"since\" : 1776126582771000\n        cluster.go:125: \t}\n        cluster.go:125: ]'\n        cluster.go:125: + printf 'sysext list:\\n%s\\nsysext status:\\n%s\\n' '[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"containerd-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/containerd-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776108218939451\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"docker-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/docker-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776108247798319\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"oem-test\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/oem-test.raw\",\n        cluster.go:125: \t\t\"time\" : 1776126551832074\n        cluster.go:125: \t}\n        cluster.go:125: ]' '[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/opt\",\n        cluster.go:125: \t\t\"extensions\" : \"none\",\n        cluster.go:125: \t\t\"since\" : null\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/usr\",\n        cluster.go:125: \t\t\"extensions\" : [\n        cluster.go:125: \t\t\t\"containerd-flatcar\",\n        cluster.go:125: \t\t\t\"docker-flatcar\",\n        cluster.go:125: \t\t\t\"oem-test\"\n        cluster.go:125: \t\t],\n        cluster.go:125: \t\t\"since\" : 1776126582771000\n        cluster.go:125: \t}\n        cluster.go:125: ]'\n        cluster.go:125: ++ jq '.[] | select(.name == \"oem-test\")'\n        cluster.go:125: + list_oem_test='{\n        cluster.go:125:   \"name\": \"oem-test\",\n        cluster.go:125:   \"type\": \"raw\",\n        cluster.go:125:   \"path\": \"/etc/extensions/oem-test.raw\",\n        cluster.go:125:   \"time\": 1776126551832074\n        cluster.go:125: }'\n        cluster.go:125: + [[ -z {\n        cluster.go:125:   \"name\": \"oem-test\",\n        cluster.go:125:   \"type\": \"raw\",\n        cluster.go:125:   \"path\": \"/etc/extensions/oem-test.raw\",\n        cluster.go:125:   \"time\": 1776126551832074\n        cluster.go:125: } ]]\n        cluster.go:125: ++ jq --raw-output .type\n        cluster.go:125: + oem_test_type=raw\n        cluster.go:125: + [[ raw != \\r\\a\\w ]]\n        cluster.go:125: ++ jq --raw-output .path\n        cluster.go:125: + oem_test_path=/etc/extensions/oem-test.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + oem_test_real_path=/oem/sysext/oem-test-4081.3.7.raw\n        cluster.go:125: + [[ /etc/extensions/oem-test.raw != \\/\\o\\e\\m\\/\\s\\y\\s\\e\\x\\t\\/\\o\\e\\m\\-\\t\\e\\s\\t\\-\\4\\0\\8\\1\\.\\3\\.\\7\\.\\r\\a\\w ]]\n        cluster.go:125: + [[ /etc/extensions/oem-test.raw != \\/\\e\\t\\c\\/\\e\\x\\t\\e\\n\\s\\i\\o\\n\\s\\/\\o\\e\\m\\-\\t\\e\\s\\t\\.\\r\\a\\w ]]\n        cluster.go:125: ++ jq '.[] | select(.hierarchy == \"/usr\")'\n        cluster.go:125: + status_usr='{\n        cluster.go:125:   \"hierarchy\": \"/usr\",\n        cluster.go:125:   \"extensions\": [\n        cluster.go:125:     \"containerd-flatcar\",\n        cluster.go:125:     \"docker-flatcar\",\n        cluster.go:125:     \"oem-test\"\n        cluster.go:125:   ],\n        cluster.go:125:   \"since\": 1776126582771000\n        cluster.go:125: }'\n        cluster.go:125: + [[ -z {\n        cluster.go:125:   \"hierarchy\": \"/usr\",\n        cluster.go:125:   \"extensions\": [\n        cluster.go:125:     \"containerd-flatcar\",\n        cluster.go:125:     \"docker-flatcar\",\n        cluster.go:125:     \"oem-test\"\n        cluster.go:125:   ],\n        cluster.go:125:   \"since\": 1776126582771000\n        cluster.go:125: } ]]\n        cluster.go:125: ++ jq --raw-output '.extensions[] | select(. == \"oem-test\")'\n        cluster.go:125: + status_usr_extensions_oem_test=oem-test\n        cluster.go:125: + [[ oem-test != \\o\\e\\m\\-\\t\\e\\s\\t ]]\n        cluster.go:125: + f=/usr/share/flatcar-sysext-kola-test/file\n        cluster.go:125: + [[ ! -e /usr/share/flatcar-sysext-kola-test/file ]]\n        cluster.go:125: ++ cat /usr/share/flatcar-sysext-kola-test/file\n        cluster.go:125: + got=4081.3.7\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-13-1915\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + ex=4081.3.7\n        cluster.go:125: + [[ 4081.3.7 != \\4\\0\\8\\1\\.\\3\\.\\7 ]]\n"},{"name":"kubeadm.v1.33.8.flannel.base/NFS_deployment","result":"PASS","duration":127408433980,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.33.8.flannel.base","result":"PASS","duration":971238177046,"output":"        cluster.go:125: I0414 00:15:35.429621    1791 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0414 00:16:03.831231    2029 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.10\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0414 00:16:04.342359    2029 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.74]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.505400228s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.74:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 11.604860993s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 32.494861036s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 1m24.063942077s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 946ogu.an0czjfzxn1yhj6x\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.74:6443 --token 946ogu.an0czjfzxn1yhj6x \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:cab005005fd2d070c38025dafcf6fe1530b4354d5fd41d115f49cb15aa19b859 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0414 00:18:52.726760    1618 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.flannel.base/node_readiness (258.19s)\n    --- PASS: kubeadm.v1.33.8.flannel.base/nginx_deployment (176.13s)\n    --- PASS: kubeadm.v1.33.8.flannel.base/NFS_deployment (127.41s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.35.1.calico.base/node_readiness","result":"PASS","duration":989843794952,"output":""},{"name":"kubeadm.v1.35.1.cilium.base/node_readiness","result":"PASS","duration":11483542633,"output":""},{"name":"kubeadm.v1.35.1.cilium.base/nginx_deployment","result":"PASS","duration":12788875445,"output":""},{"name":"kubeadm.v1.35.1.cilium.base/NFS_deployment","result":"PASS","duration":14601595822,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"cl.flannel.udp","result":"FAIL","duration":2675225476554,"output":"        harness.go:646: Cluster failed starting machines: machine \"8b4ec69c-59e5-4be6-8a77-79d055f4a29c\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.26:22: connect: no route to host\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-udev-settl…\u001b[0m To Complete Device Initialization.\r\r) on machine 8b4ec69c-59e5-4be6-8a77-79d055f4a29c console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 8b4ec69c-59e5-4be6-8a77-79d055f4a29c console\n"},{"name":"kubeadm.v1.35.1.cilium.base/IPSec_encryption","result":"PASS","duration":22900193087,"output":"        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"kubeadm.v1.35.1.cilium.base","result":"PASS","duration":855368060248,"output":"        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.13.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.6-0\n        cluster.go:125: [init] Using Kubernetes version: v1.35.3\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0414 00:20:21.211310    2181 checks.go:906] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.81]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.126161962s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.81:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 13.527951995s\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 14.143653822s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 8m41.537967766s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 2iu0l7.sg2te6dsw3cj7grr\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.81:6443 --token 2iu0l7.sg2te6dsw3cj7grr \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:a91388ef84072892d6e533aee62fde67dcf095aac3863ddb49d72504aef29a88 \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0414 00:30:13.113618    1652 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.35.1.cilium.base/node_readiness (11.48s)\n    --- PASS: kubeadm.v1.35.1.cilium.base/nginx_deployment (12.79s)\n    --- PASS: kubeadm.v1.35.1.cilium.base/NFS_deployment (14.60s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n    --- PASS: kubeadm.v1.35.1.cilium.base/IPSec_encryption (22.90s)\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"kubeadm.v1.35.1.calico.base/nginx_deployment","result":"PASS","duration":114205544528,"output":""},{"name":"kubeadm.v1.35.1.calico.base/NFS_deployment","result":"PASS","duration":18322546193,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.35.1.calico.base","result":"PASS","duration":1960971304740,"output":"        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.3\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.13.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.6-0\n        cluster.go:125: [init] Using Kubernetes version: v1.35.3\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0414 00:11:55.565991    2194 checks.go:906] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.46]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.017794159s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.46:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 5.086798447s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 10.876966351s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 24.011857275s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 67ho99.mdkp5shuypqmhena\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.46:6443 --token 67ho99.mdkp5shuypqmhena \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:947b4ef64c62ecd2afb3ee07d7402e6816d7cae7d5afcdd763ae151f1c408fff \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: installation.operator.tigera.io/default created\n        cluster.go:125: apiserver.operator.tigera.io/default created\n        cluster.go:125: goldmane.operator.tigera.io/default created\n        cluster.go:125: whisker.operator.tigera.io/default created\n        cluster.go:125: W0414 00:13:36.469057    1673 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.35.1.calico.base/node_readiness (989.84s)\n    --- PASS: kubeadm.v1.35.1.calico.base/nginx_deployment (114.21s)\n    --- PASS: kubeadm.v1.35.1.calico.base/NFS_deployment (18.32s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"}],"result":"FAIL","platform":"qemu","version":"4081.3.7"}
