{"tests":[{"name":"cl.ignition.v1.btrfsroot","result":"PASS","duration":22645180894,"output":""},{"name":"kubeadm.v1.34.4.flannel.base/node_readiness","result":"PASS","duration":16557507790,"output":""},{"name":"kubeadm.v1.34.4.flannel.base/nginx_deployment","result":"PASS","duration":11469119347,"output":""},{"name":"cl.sysext.boot","result":"PASS","duration":99934765416,"output":"        cluster.go:125: + sudo systemctl mask --now ensure-sysext\n        cluster.go:125: Created symlink /etc/systemd/system/ensure-sysext.service → /dev/null.\n        cluster.go:125: + sudo systemctl mask --now systemd-sysext\n        cluster.go:125: Created symlink /etc/systemd/system/systemd-sysext.service → /dev/null.\n        cluster.go:125: + sudo mkdir -p /etc/flatcar/sysext /etc/flatcar/oem-sysext /oem/sysext /etc/extensions\n        cluster.go:125: + echo ID=test\n        cluster.go:125: + sudo tee /oem/oem-release\n        cluster.go:125: + echo myext\n        cluster.go:125: + sudo tee /etc/flatcar/enabled-sysext.conf\n        cluster.go:125: + sudo touch /oem/sysext/active-oem-test /etc/flatcar/oem-sysext/oem-test-4081.3.7.raw /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw /etc/flatcar/sysext/flatcar-myext-1.2.3.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /etc/extensions/oem-test.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/sysext/flatcar-myext-1.2.3.raw /etc/extensions/flatcar-myext.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4081.3.7.raw = /oem/sysext/oem-test-4081.3.7.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw = /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw ']'\n        cluster.go:125: + sudo mv /oem/sysext/oem-test-4081.3.7.raw /etc/flatcar/oem-sysext/\n        cluster.go:125: + sudo mv /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /oem/sysext/\n        cluster.go:125: + sudo ln -fs /oem/sysext/oem-test-1.2.3.raw /etc/extensions/oem-test.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/sysext/flatcar-myext-1.2.3.raw /etc/extensions/flatcar-myext.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4081.3.7.raw = /oem/sysext/oem-test-4081.3.7.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw = /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw ']'\n        cluster.go:125: + '[' -e /etc/flatcar/oem-sysext/oem-test-1.2.3.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4081.3.7.raw = /oem/sysext/oem-test-4081.3.7.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw = /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw ']'\n"},{"name":"kubeadm.v1.34.4.flannel.base/NFS_deployment","result":"PASS","duration":17475950315,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.34.4.flannel.base","result":"PASS","duration":141559370224,"output":"        cluster.go:125: I0421 10:06:13.513847    1763 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0421 10:06:24.351754    1983 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:06:24.559072    1983 checks.go:827] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.5]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.004356141s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.5:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.786355263s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 2.773540585s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 4.502594127s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: grtx06.4qjify5ci6qday8x\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.5:6443 --token grtx06.4qjify5ci6qday8x \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:c3a778aaf763fb8c32a3fdfab194bf397d837da2b555d63ea77c437ff37f65c3 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0421 10:07:01.076932    1619 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.34.4.flannel.base/node_readiness (16.56s)\n    --- PASS: kubeadm.v1.34.4.flannel.base/nginx_deployment (11.47s)\n    --- PASS: kubeadm.v1.34.4.flannel.base/NFS_deployment (17.48s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.34.4.calico.base/node_readiness","result":"PASS","duration":27875200333,"output":""},{"name":"kubeadm.v1.34.4.calico.base/nginx_deployment","result":"PASS","duration":6324685750,"output":""},{"name":"cl.tpm.root-cryptenroll-pcr-withupdate","result":"FAIL","duration":171567739354,"output":"        cluster.go:125: Warning: keyslot operation could fail as it requires more than available memory.\n        cluster.go:125: New TPM2 token enrolled as key slot 1.\n        cluster.go:125: Wiped slot 2.\n        tpm.go:354: could not reboot machine: machine \"d0caaff7-073f-42e7-a2e3-93876fd630c6\" failed basic checks: some systemd units failed:\n● ldconfig.service loaded failed failed Rebuild Dynamic Linker Cache\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39mldconfig.service\u001b[0m - Rebuild Dynamic Linker Cache.\r\r) on machine d0caaff7-073f-42e7-a2e3-93876fd630c6 console\n"},{"name":"kubeadm.v1.34.4.calico.base/NFS_deployment","result":"PASS","duration":19413915104,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.34.4.calico.base","result":"PASS","duration":164978103956,"output":"        cluster.go:125: I0421 10:08:37.200231    1798 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0421 10:08:47.479463    2018 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:08:47.599622    2018 checks.go:827] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.11]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.006051181s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.11:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.515222271s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 5.008573728s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 6.502547789s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: xpeu23.njck2h5jr3lqf1qg\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.11:6443 --token xpeu23.njck2h5jr3lqf1qg \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:42da36df8abfee8777861a10f9fd45eed0af5ffa747710cd0ced0b9db3bb2312 \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: installation.operator.tigera.io/default created\n        cluster.go:125: apiserver.operator.tigera.io/default created\n        cluster.go:125: goldmane.operator.tigera.io/default created\n        cluster.go:125: whisker.operator.tigera.io/default created\n        cluster.go:125: W0421 10:09:38.995308    1644 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.34.4.calico.base/node_readiness (27.88s)\n    --- PASS: kubeadm.v1.34.4.calico.base/nginx_deployment (6.32s)\n    --- PASS: kubeadm.v1.34.4.calico.base/NFS_deployment (19.41s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"cl.ignition.symlink","result":"PASS","duration":22336983258,"output":"        cluster.go:152: + readlink /etc/localtime\n"},{"name":"cl.ignition.v2.users","result":"PASS","duration":21849445215,"output":""},{"name":"systemd.journal.user","result":"PASS","duration":21235272133,"output":""},{"name":"coreos.locksmith.reboot","result":"PASS","duration":345430586719,"output":""},{"name":"cl.verity/verify","result":"PASS","duration":3626654722,"output":"        cluster.go:125: Success\n"},{"name":"cl.toolbox.dnf-install","result":"PASS","duration":36801597598,"output":""},{"name":"cl.osreset.ignition-rerun","result":"PASS","duration":45294902363,"output":""},{"name":"cl.ignition.oem.indirect","result":"PASS","duration":44120216479,"output":""},{"name":"cl.ignition.partition_on_boot_disk","result":"PASS","duration":21945662682,"output":"        filesystem.go:487: lsblk -f:\nNAME    FSTYPE   FSVER LABEL      UUID                                 FSAVAIL FSUSE% MOUNTPOINTS\nloop2   squashfs 4.0                                                                  \nloop3   squashfs 4.0                                                                  \nsr0                                                                                   \nvda                                                                                   \n|-vda1  vfat     FAT32 EFI-SYSTEM EAC4-8906                                           \n|-vda2                                                                                \n|-vda3  btrfs                     4627a20b-c3ad-458e-a05a-90623574a539                \n| `-usr btrfs                     4627a20b-c3ad-458e-a05a-90623574a539                /usr\n|-vda4                                                                                \n|-vda6  btrfs          OEM        855d7a31-c001-47db-a073-492800715453    123M     1% /oem\n|-vda7                                                                                \n|-vda9  ext4     1.0   ROOT       fd5e5f40-ad85-46ea-abb5-3cc3d4cd8af5    5.9G     0% /\n`-vda10 xfs            VAR        d68dc32b-a62e-417a-a1a0-a401ac4d5b56    5.2G     3% /var\n        filesystem.go:489: findmnt:\nTARGET                        SOURCE     FSTYPE     OPTIONS\n/                             /dev/vda9  ext4       rw,relatime,seclabel\n|-/boot                       systemd-1  autofs     rw,relatime,fd=33,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=7225\n|-/media                      tmpfs      tmpfs      rw,nosuid,nodev,noexec,relatime,seclabel\n|-/tmp                        tmpfs      tmpfs      rw,nosuid,nodev,seclabel,size=1236924k,nr_inodes=1048576\n|-/dev                        devtmpfs   devtmpfs   rw,nosuid,seclabel,size=4096k,nr_inodes=303641,mode=755\n| |-/dev/hugepages            hugetlbfs  hugetlbfs  rw,nosuid,nodev,relatime,seclabel,pagesize=2M\n| |-/dev/mqueue               mqueue     mqueue     rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/dev/shm                  tmpfs      tmpfs      rw,nosuid,nodev,seclabel\n| `-/dev/pts                  devpts     devpts     rw,nosuid,noexec,relatime,seclabel,gid=5,mode=620,ptmxmode=000\n|-/sys                        sysfs      sysfs      rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/fs/selinux           selinuxfs  selinuxfs  rw,nosuid,noexec,relatime\n| |-/sys/kernel/debug         debugfs    debugfs    rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/kernel/tracing       tracefs    tracefs    rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/kernel/security      securityfs securityfs rw,nosuid,nodev,noexec,relatime\n| |-/sys/fs/cgroup            cgroup2    cgroup2    rw,nosuid,nodev,noexec,relatime,seclabel,nsdelegate,memory_recursiveprot\n| |-/sys/fs/pstore            pstore     pstore     rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/firmware/efi/efivars efivarfs   efivarfs   rw,nosuid,nodev,noexec,relatime\n| |-/sys/fs/bpf               bpf        bpf        rw,nosuid,nodev,noexec,relatime,mode=700\n| |-/sys/fs/fuse/connections  fusectl    fusectl    rw,nosuid,nodev,noexec,relatime\n| `-/sys/kernel/config        configfs   configfs   rw,nosuid,nodev,noexec,relatime\n|-/proc                       proc       proc       rw,nosuid,nodev,noexec,relatime\n| `-/proc/sys/fs/binfmt_misc  systemd-1  autofs     rw,relatime,fd=34,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=7229\n|-/run                        tmpfs      tmpfs      rw,nosuid,nodev,seclabel,size=494772k,nr_inodes=819200,mode=755\n| `-/run/user/500             tmpfs      tmpfs      rw,nosuid,nodev,relatime,seclabel,size=247384k,nr_inodes=61846,mode=700,uid=500,gid=500\n|-/var                        /dev/vda10 xfs        rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota\n|-/usr                        /dev/dm-0  btrfs      ro,relatime,seclabel,rescue=nologreplay,space_cache=v2,subvolid=5,subvol=/\n| `-/usr                      sysext     overlay    ro,nodev,relatime,seclabel,lowerdir=/run/systemd/sysext/meta/usr:/run/systemd/sysext/extensions/docker-flatcar/usr:/run/systemd/sysext/extensions/containerd-flatcar/usr:/usr\n|-/etc                        overlay    overlay    rw,noatime,seclabel,lowerdir=/sysroot/usr/share/flatcar/etc,upperdir=/sysroot/etc,workdir=/sysroot/.etc-work,uuid=on,metacopy=off\n`-/oem                        /dev/vda6  btrfs      rw,nodev,relatime,seclabel,discard=async,space_cache=v2,subvolid=5,subvol=/\n        cluster.go:152: + findmnt /var\n        cluster.go:152: + cat /var/hello\n"},{"name":"cl.ignition.v1.users","result":"PASS","duration":22905285368,"output":""},{"name":"cl.users.shells","result":"PASS","duration":22332457943,"output":""},{"name":"cl.tang.root","result":"PASS","duration":75366018564,"output":"        tang.go:148: Started tang on 10.0.0.1:34801\n"},{"name":"cl.verity/corruption","result":"PASS","duration":168865186398,"output":""},{"name":"cl.verity","result":"PASS","duration":195828468511,"output":"    --- PASS: cl.verity/verify (3.63s)\n            cluster.go:125: Success\n    --- PASS: cl.verity/corruption (168.87s)\n"},{"name":"kubeadm.v1.33.8.flannel.cgroupv1.base/node_readiness","result":"PASS","duration":10790793776,"output":""},{"name":"cl.ignition.oem.indirect.new","result":"PASS","duration":49637368225,"output":""},{"name":"kubeadm.v1.33.8.flannel.cgroupv1.base/nginx_deployment","result":"PASS","duration":26899013809,"output":""},{"name":"cl.ignition.v1.ext4root","result":"PASS","duration":30363895907,"output":""},{"name":"kubeadm.v1.33.8.flannel.cgroupv1.base/NFS_deployment","result":"PASS","duration":11843561659,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.33.8.flannel.cgroupv1.base","result":"PASS","duration":168150507261,"output":"        cluster.go:125: I0421 10:13:44.932770    1884 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0421 10:13:58.191440    2108 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:13:58.367465    2108 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.25]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 502.220519ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.25:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.506345812s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 2.053891137s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 4.005802515s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: izll99.j963h1jpck3fw1h3\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.25:6443 --token izll99.j963h1jpck3fw1h3 \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:c437b2ba6867adef75037813cb6d7ad0f04daf7867761a6094300e107cfb0dbc \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0421 10:14:35.668629    1743 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.flannel.cgroupv1.base/node_readiness (10.79s)\n    --- PASS: kubeadm.v1.33.8.flannel.cgroupv1.base/nginx_deployment (26.90s)\n    --- PASS: kubeadm.v1.33.8.flannel.cgroupv1.base/NFS_deployment (11.84s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"cl.omaha.ping","result":"PASS","duration":22446412084,"output":""},{"name":"coreos.ignition.sethostname","result":"PASS","duration":22257761631,"output":""},{"name":"docker.base/docker-info","result":"PASS","duration":663839713,"output":""},{"name":"docker.base/resources","result":"PASS","duration":2589646507,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.20MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:aa311dd39c1e07c06f003de6fa4287b2cd74698bf581f4b16487babc4075173d done\n        cluster.go:125: #5 naming to docker.io/library/sleep done\n        cluster.go:125: #5 DONE 0.0s\n        cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n        cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n"},{"name":"cl.overlay.cleanup","result":"PASS","duration":34052174389,"output":""},{"name":"cl.disk.raid1.root","result":"PASS","duration":43781688109,"output":""},{"name":"cl.disk.raid0.data","result":"PASS","duration":43037893767,"output":""},{"name":"devcontainer.systemd-nspawn","result":"PASS","duration":212507826408,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + ARCH=amd64\n        cluster.go:125: + VERSION=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + EXPECTED_VERSION=4081.3.7\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-19471\n        cluster.go:125: + USR_SRC_DIR=/home/core/dev-container-workdir-19471/src\n        cluster.go:125: + VAR_TMP_DIR=/home/core/dev-container-workdir-19471/tmp\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-19471/src /home/core/dev-container-workdir-19471/tmp\n        cluster.go:125: + sudo systemd-nspawn --console=pipe --setenv=PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs --setenv=EXPECTED_VERSION=4081.3.7 --bind-ro=/lib/modules --bind-ro=/home/core/dev-container-script --bind=/home/core/dev-container-workdir-19471/src:/usr/src --bind=/home/core/dev-container-workdir-19471/tmp:/var/tmp --image=flatcar_developer_container.bin --machine=flatcar-developer-container /bin/bash /home/core/dev-container-script\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + [[ 4081.3.7 != \\4\\0\\8\\1\\.\\3\\.\\7 ]]\n        cluster.go:125: + export PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + export 'FEATURES=-ipc-sandbox -network-sandbox'\n        cluster.go:125: + FEATURES='-ipc-sandbox -network-sandbox'\n        cluster.go:125: + emerge-gitclone\n        cluster.go:125: Cloning into '/var/lib/portage/scripts'...\n        cluster.go:125: Note: switching to 'lts-4081.3.7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at a4738dc8a7 New version: lts-4081.3.7\n        cluster.go:125: + emerge --getbinpkg --verbose coreos-sources\n        cluster.go:125: \n        cluster.go:125: !!! Your current profile is deprecated and not supported anymore.\n        cluster.go:125: !!! Use eselect profile to update your profile.\n        cluster.go:125: !!! Please upgrade to the following profile if possible:\n        cluster.go:125: \n        cluster.go:125:         default/linux/amd64/23.0/split-usr/no-multilib/hardened\n        cluster.go:125: \n        cluster.go:125: To upgrade do the following steps:\n        cluster.go:125: \n        cluster.go:125: A profile upgrade to version 23.0 is available for your architecture.\n        cluster.go:125: The new 23.0 profiles enable some toolchain hardening features and \n        cluster.go:125: performance enhancements by default, and standardize settings.\n        cluster.go:125: You can find the list of changes on the wiki tracking page [1].\n        cluster.go:125: \n        cluster.go:125: Upgrade instructions\n        cluster.go:125: \n        cluster.go:125: Note 1: If you have manually changed your CHOST to a value different from \n        cluster.go:125: what the stages and profiles set, you may have to do that in the future too.\n        cluster.go:125: In that case you should know what you are doing, hopefully; please read the \n        cluster.go:125: instructions with a critical eye then.\n        cluster.go:125: \n        cluster.go:125: Note 2: In case you are already familiar with binary packages, you should be\n        cluster.go:125: able to add \"--getbinpkg\" to the emerge calls to speed things up.\n        cluster.go:125: The use of binary packages is completely optional though, and also not\n        cluster.go:125: as much tested as the source-based upgrade path yet.\n        cluster.go:125: \n        cluster.go:125: 1. Ensure your system backups are up to date. Please also update\n        cluster.go:125:    your system fully and depclean before proceeding.\n        cluster.go:125:    glibc older than 2.36 and musl older than 1.2.4 is not supported anymore.\n        cluster.go:125: \n        cluster.go:125: 2. If you are still using one of the long-deprecated amd64 17.0 profiles \n        cluster.go:125:    (other than x32 or musl), then first complete the migration to the \n        cluster.go:125:    corresponding 17.1 profile. Instructions can be found at [3].\n        cluster.go:125:    \n        cluster.go:125: 3. If you are currently using systemd in a split-usr configuration, then first \n        cluster.go:125:    complete the migration to the corresponding merged-usr profile of the \n        cluster.go:125:    same profile version. Details on how to do this can be found in the news \n        cluster.go:125:    item [4].\n        cluster.go:125:    If you are currently using openrc, migrate to 23.0 first, keeping your disk\n        cluster.go:125:    layout. If you want to move from split-usr to merged-usr, do that afterwards.\n        cluster.go:125: \n        cluster.go:125: 4. Run \"emerge --info\" and note down the value of the CHOST variable.\n        cluster.go:125: \n        cluster.go:125: 5. Edit /etc/portage/make.conf; if there is a line defining the CHOST variable,\n        cluster.go:125:    remove it. Also delete all lines defining CHOST_... variables.\n        cluster.go:125: \n        cluster.go:125: 6. Select the 23.0 profile corresponding to your current profile, either using\n        cluster.go:125:    \"eselect profile\" or by manually setting the profile symlink.\n        cluster.go:125:    Note that old profiles are by default split-usr and the 23.0 profiles by\n        cluster.go:125:    default merged-usr. Do NOT change directory scheme now, since this will\n        cluster.go:125:    mess up your system! \n        cluster.go:125:    Instead, make sure that the new profile has the same property: for example, \n        cluster.go:125:    OLD default/linux/amd64/17.1  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/split-usr\n        cluster.go:125:              (added \"split-usr\")\n        cluster.go:125:    OLD default/linux/amd64/17.1/systemd/merged-usr  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/systemd\n        cluster.go:125:              (removed \"merged-usr\")\n        cluster.go:125:    A detailed table of the upgrade paths can be found at [5]. Please consult it.\n        cluster.go:125:    In some cases (hppa, x86) the table will tell you to pick between two choices. \n        cluster.go:125:    What you need should be obvious from your *old* CHOST value (from step 4).\n        cluster.go:125: \n        cluster.go:125: 7. Delete the contents of your binary package cache at ${PKGDIR}\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 8. In the file or directory /etc/portage/binrepos.conf (if existing), update\n        cluster.go:125:    the URI in all configuration such that they point to 23.0 profile binhost \n        cluster.go:125:    directories. The exact paths can be found in the table at [5], too.\n        cluster.go:125: \n        cluster.go:125: 9. Rebuild or reinstall from binary (if available) the following packages in\n        cluster.go:125:    this order, with the same version as already active:\n        cluster.go:125:      emerge --ask --oneshot sys-devel/binutils\n        cluster.go:125:    (you may have to run binutils-config and re-select your binutils now)\n        cluster.go:125:      emerge --ask --oneshot sys-devel/gcc\n        cluster.go:125:    (IMPORTANT: If this command wants to rebuild glibc first, do *not* let it do \n        cluster.go:125:     that; instead, abort and try again with --nodeps added to the command line.)\n        cluster.go:125:    (you may have to run gcc-config and re-select your gcc now)\n        cluster.go:125:    and the C library, i.e. for glibc-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/glibc\n        cluster.go:125:    or for musl-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/musl\n        cluster.go:125: \n        cluster.go:125: 10. Re-run \"emerge --info\" and check if CHOST has changed compared to step 4.\n        cluster.go:125: \n        cluster.go:125: If the CHOST has NOT changed, skip to step 13 (env-update). Otherwise, \n        cluster.go:125: \n        cluster.go:125: 11. Recheck with binutils-config and gcc-config that valid installed versions\n        cluster.go:125:    of binutils and gcc are selected.\n        cluster.go:125: \n        cluster.go:125: 12. Check /etc/env.d, /etc/env.d/binutils, and /etc/env.d/gcc for files that\n        cluster.go:125:    refer to the *OLD* CHOST value, and remove them. \n        cluster.go:125:    Examples how to do this can be found in the similar procedure at [6].\n        cluster.go:125: \n        cluster.go:125: 13. Run env-update \u0026\u0026 source /etc/profile\n        cluster.go:125: \n        cluster.go:125: 14. Re-emerge libtool:\n        cluster.go:125:    emerge --ask --oneshot libtool\n        cluster.go:125: \n        cluster.go:125: 15. Just for safety, delete the contents of your binary package cache at \n        cluster.go:125:     ${PKGDIR} again:\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 16. Rebuild world:\n        cluster.go:125:    emerge --ask --emptytree @world\n        cluster.go:125: \n        cluster.go:125: [1] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_transition\n        cluster.go:125: [2] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_timeline\n        cluster.go:125: [3] https://www.gentoo.org/support/news-items/2019-06-05-amd64-17-1-profiles-are-now-stable.html\n        cluster.go:125: [4] https://www.gentoo.org/support/news-items/2022-12-01-systemd-usrmerge.html\n        cluster.go:125: [5] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_update_table\n        cluster.go:125: [6] https://wiki.gentoo.org/wiki/Changing_the_CHOST_variable#Verifying_things_work\n        cluster.go:125: \n        cluster.go:125: \n        cluster.go:125: + zcat /proc/config.gz\n        cluster.go:125: ++ nproc\n        cluster.go:125: + exec make -C /usr/src/linux -j4 modules_prepare V=1\n"},{"name":"docker.base/networks-reliably","result":"PASS","duration":128502248477,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 4.04MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:d56cd558c87e3b3d3c02bfa7a0ff5d4716a382b79f1f77bdbb08376a9ce4bf28 done\n        cluster.go:125: #5 naming to docker.io/library/ping done\n        cluster.go:125: #5 DONE 0.0s\n"},{"name":"docker.base/user-no-caps","result":"PASS","duration":807010330,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 5.06MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:5a6dcf1d20fea4aac9efcb7a09f254bf3d3cd7c547460700ec628a502cd91d4f done\n        cluster.go:125: #5 naming to docker.io/library/captest done\n        cluster.go:125: #5 DONE 0.0s\n"},{"name":"cl.cloudinit.multipart-mime","result":"PASS","duration":22913712809,"output":""},{"name":"docker.base/ownership","result":"PASS","duration":4220556180,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/nginx\n        cluster.go:125: 9d11882d4b58: Pulling fs layer\n        cluster.go:125: b44d445f805d: Pulling fs layer\n        cluster.go:125: 8bef7a9ebe88: Pulling fs layer\n        cluster.go:125: a0f1c91fa11d: Pulling fs layer\n        cluster.go:125: 2044bb7170df: Pulling fs layer\n        cluster.go:125: 95d77c54a9aa: Pulling fs layer\n        cluster.go:125: c13dd6bf8c10: Pulling fs layer\n        cluster.go:125: a0f1c91fa11d: Waiting\n        cluster.go:125: 2044bb7170df: Waiting\n        cluster.go:125: 95d77c54a9aa: Waiting\n        cluster.go:125: c13dd6bf8c10: Waiting\n        cluster.go:125: 8bef7a9ebe88: Verifying Checksum\n        cluster.go:125: 8bef7a9ebe88: Download complete\n        cluster.go:125: b44d445f805d: Verifying Checksum\n        cluster.go:125: b44d445f805d: Download complete\n        cluster.go:125: 9d11882d4b58: Download complete\n        cluster.go:125: a0f1c91fa11d: Verifying Checksum\n        cluster.go:125: a0f1c91fa11d: Download complete\n        cluster.go:125: 95d77c54a9aa: Verifying Checksum\n        cluster.go:125: 95d77c54a9aa: Download complete\n        cluster.go:125: 2044bb7170df: Verifying Checksum\n        cluster.go:125: 2044bb7170df: Download complete\n        cluster.go:125: c13dd6bf8c10: Verifying Checksum\n        cluster.go:125: c13dd6bf8c10: Download complete\n        cluster.go:125: 9d11882d4b58: Pull complete\n        cluster.go:125: b44d445f805d: Pull complete\n        cluster.go:125: 8bef7a9ebe88: Pull complete\n        cluster.go:125: a0f1c91fa11d: Pull complete\n        cluster.go:125: 2044bb7170df: Pull complete\n        cluster.go:125: 95d77c54a9aa: Pull complete\n        cluster.go:125: c13dd6bf8c10: Pull complete\n        cluster.go:125: Digest: sha256:6f170eb5f8ea893d94fe3c05cef69270f16614a99d3d922ce8a320efc43b8591\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"docker.base","result":"PASS","duration":158062201321,"output":"    --- PASS: docker.base/docker-info (0.66s)\n    --- PASS: docker.base/resources (2.59s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 2.20MB 0.0s done\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.0s done\n            cluster.go:125: #5 writing image sha256:aa311dd39c1e07c06f003de6fa4287b2cd74698bf581f4b16487babc4075173d done\n            cluster.go:125: #5 naming to docker.io/library/sleep done\n            cluster.go:125: #5 DONE 0.0s\n            cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n            cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n    --- PASS: docker.base/networks-reliably (128.50s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 4.04MB 0.0s done\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.0s done\n            cluster.go:125: #5 writing image sha256:d56cd558c87e3b3d3c02bfa7a0ff5d4716a382b79f1f77bdbb08376a9ce4bf28 done\n            cluster.go:125: #5 naming to docker.io/library/ping done\n            cluster.go:125: #5 DONE 0.0s\n    --- PASS: docker.base/user-no-caps (0.81s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 5.06MB 0.0s done\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.0s done\n            cluster.go:125: #5 writing image sha256:5a6dcf1d20fea4aac9efcb7a09f254bf3d3cd7c547460700ec628a502cd91d4f done\n            cluster.go:125: #5 naming to docker.io/library/captest done\n            cluster.go:125: #5 DONE 0.0s\n    --- PASS: docker.base/ownership (4.22s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/nginx\n            cluster.go:125: 9d11882d4b58: Pulling fs layer\n            cluster.go:125: b44d445f805d: Pulling fs layer\n            cluster.go:125: 8bef7a9ebe88: Pulling fs layer\n            cluster.go:125: a0f1c91fa11d: Pulling fs layer\n            cluster.go:125: 2044bb7170df: Pulling fs layer\n            cluster.go:125: 95d77c54a9aa: Pulling fs layer\n            cluster.go:125: c13dd6bf8c10: Pulling fs layer\n            cluster.go:125: a0f1c91fa11d: Waiting\n            cluster.go:125: 2044bb7170df: Waiting\n            cluster.go:125: 95d77c54a9aa: Waiting\n            cluster.go:125: c13dd6bf8c10: Waiting\n            cluster.go:125: 8bef7a9ebe88: Verifying Checksum\n            cluster.go:125: 8bef7a9ebe88: Download complete\n            cluster.go:125: b44d445f805d: Verifying Checksum\n            cluster.go:125: b44d445f805d: Download complete\n            cluster.go:125: 9d11882d4b58: Download complete\n            cluster.go:125: a0f1c91fa11d: Verifying Checksum\n            cluster.go:125: a0f1c91fa11d: Download complete\n            cluster.go:125: 95d77c54a9aa: Verifying Checksum\n            cluster.go:125: 95d77c54a9aa: Download complete\n            cluster.go:125: 2044bb7170df: Verifying Checksum\n            cluster.go:125: 2044bb7170df: Download complete\n            cluster.go:125: c13dd6bf8c10: Verifying Checksum\n            cluster.go:125: c13dd6bf8c10: Download complete\n            cluster.go:125: 9d11882d4b58: Pull complete\n            cluster.go:125: b44d445f805d: Pull complete\n            cluster.go:125: 8bef7a9ebe88: Pull complete\n            cluster.go:125: a0f1c91fa11d: Pull complete\n            cluster.go:125: 2044bb7170df: Pull complete\n            cluster.go:125: 95d77c54a9aa: Pull complete\n            cluster.go:125: c13dd6bf8c10: Pull complete\n            cluster.go:125: Digest: sha256:6f170eb5f8ea893d94fe3c05cef69270f16614a99d3d922ce8a320efc43b8591\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"coreos.ignition.security.tls","result":"PASS","duration":41474498765,"output":""},{"name":"cl.network.iptables","result":"PASS","duration":21352667389,"output":"        cluster.go:152: + sudo nft --json list ruleset | jq '.nftables[] | select(.rule) | .rule.expr[0].match.right'\n"},{"name":"cl.ignition.kargs","result":"PASS","duration":31514381881,"output":"        cluster.go:152: + cat /proc/cmdline\n"},{"name":"bpf.ig/ig","result":"PASS","duration":31000231465,"output":"        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n        cluster.go:125: + docker info\n        cluster.go:125: + docker ps\n        cluster.go:125: + docker images\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n        cluster.go:125: + dig kinvolk.io\n        cluster.go:125: + dig flatcar.org\n        cluster.go:125: + dig stable.release.flatcar-linux.net\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"bpf.ig","result":"PASS","duration":33130261126,"output":"    --- PASS: bpf.ig/ig (31.00s)\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n            cluster.go:125: + docker info\n            cluster.go:125: + docker ps\n            cluster.go:125: + docker images\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n            cluster.go:125: + dig kinvolk.io\n            cluster.go:125: + dig flatcar.org\n            cluster.go:125: + dig stable.release.flatcar-linux.net\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"cl.ignition.oem.regular.new","result":"PASS","duration":43663229094,"output":""},{"name":"cl.network.wireguard","result":"PASS","duration":22382052364,"output":"        cluster.go:152: + ip --json address show kv0 | jq -r '.[] | .addr_info | .[] | select( .family == \"inet\") | .local'\n"},{"name":"docker.lib-coreos-dockerd-compat/docker-info","result":"PASS","duration":1656654257,"output":""},{"name":"docker.lib-coreos-dockerd-compat/resources","result":"PASS","duration":5840046299,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.20MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:c8df75e8fc4864763c44937b91fbf5c3f4c3cf1b9469666729b6a50bdd0f77c6 done\n        cluster.go:125: #5 naming to docker.io/library/sleep done\n        cluster.go:125: #5 DONE 0.0s\n        cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n        cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n"},{"name":"docker.lib-coreos-dockerd-compat/networks-reliably","result":"PASS","duration":134163077395,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 4.04MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:45fdb3ba87acc2b0a0f7f284bddc57cc957d1e4211bc02bda950747cf96e95ef done\n        cluster.go:125: #5 naming to docker.io/library/ping done\n        cluster.go:125: #5 DONE 0.0s\n"},{"name":"docker.lib-coreos-dockerd-compat/user-no-caps","result":"PASS","duration":993959683,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 5.06MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:966e9eb4f508452a21a07aed95daff492ba7e512bb50e81a71833895f08c8e00 done\n        cluster.go:125: #5 naming to docker.io/library/captest done\n        cluster.go:125: #5 DONE 0.0s\n"},{"name":"docker.lib-coreos-dockerd-compat/ownership","result":"PASS","duration":4067512425,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/nginx\n        cluster.go:125: 9d11882d4b58: Pulling fs layer\n        cluster.go:125: b44d445f805d: Pulling fs layer\n        cluster.go:125: 8bef7a9ebe88: Pulling fs layer\n        cluster.go:125: a0f1c91fa11d: Pulling fs layer\n        cluster.go:125: 2044bb7170df: Pulling fs layer\n        cluster.go:125: 95d77c54a9aa: Pulling fs layer\n        cluster.go:125: c13dd6bf8c10: Pulling fs layer\n        cluster.go:125: a0f1c91fa11d: Waiting\n        cluster.go:125: 2044bb7170df: Waiting\n        cluster.go:125: 95d77c54a9aa: Waiting\n        cluster.go:125: c13dd6bf8c10: Waiting\n        cluster.go:125: 8bef7a9ebe88: Verifying Checksum\n        cluster.go:125: 8bef7a9ebe88: Download complete\n        cluster.go:125: 9d11882d4b58: Verifying Checksum\n        cluster.go:125: 9d11882d4b58: Download complete\n        cluster.go:125: b44d445f805d: Verifying Checksum\n        cluster.go:125: b44d445f805d: Download complete\n        cluster.go:125: a0f1c91fa11d: Download complete\n        cluster.go:125: 2044bb7170df: Download complete\n        cluster.go:125: 95d77c54a9aa: Verifying Checksum\n        cluster.go:125: 95d77c54a9aa: Download complete\n        cluster.go:125: c13dd6bf8c10: Verifying Checksum\n        cluster.go:125: c13dd6bf8c10: Download complete\n        cluster.go:125: 9d11882d4b58: Pull complete\n        cluster.go:125: b44d445f805d: Pull complete\n        cluster.go:125: 8bef7a9ebe88: Pull complete\n        cluster.go:125: a0f1c91fa11d: Pull complete\n        cluster.go:125: 2044bb7170df: Pull complete\n        cluster.go:125: 95d77c54a9aa: Pull complete\n        cluster.go:125: c13dd6bf8c10: Pull complete\n        cluster.go:125: Digest: sha256:6f170eb5f8ea893d94fe3c05cef69270f16614a99d3d922ce8a320efc43b8591\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"docker.lib-coreos-dockerd-compat","result":"PASS","duration":169083623776,"output":"    --- PASS: docker.lib-coreos-dockerd-compat/docker-info (1.66s)\n    --- PASS: docker.lib-coreos-dockerd-compat/resources (5.84s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 2.20MB 0.0s done\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.0s done\n            cluster.go:125: #5 writing image sha256:c8df75e8fc4864763c44937b91fbf5c3f4c3cf1b9469666729b6a50bdd0f77c6 done\n            cluster.go:125: #5 naming to docker.io/library/sleep done\n            cluster.go:125: #5 DONE 0.0s\n            cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n            cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n    --- PASS: docker.lib-coreos-dockerd-compat/networks-reliably (134.16s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 4.04MB 0.0s done\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.0s done\n            cluster.go:125: #5 writing image sha256:45fdb3ba87acc2b0a0f7f284bddc57cc957d1e4211bc02bda950747cf96e95ef done\n            cluster.go:125: #5 naming to docker.io/library/ping done\n            cluster.go:125: #5 DONE 0.0s\n    --- PASS: docker.lib-coreos-dockerd-compat/user-no-caps (0.99s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 5.06MB 0.0s done\n            cluster.go:125: #3 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.0s done\n            cluster.go:125: #5 writing image sha256:966e9eb4f508452a21a07aed95daff492ba7e512bb50e81a71833895f08c8e00 done\n            cluster.go:125: #5 naming to docker.io/library/captest done\n            cluster.go:125: #5 DONE 0.0s\n    --- PASS: docker.lib-coreos-dockerd-compat/ownership (4.07s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/nginx\n            cluster.go:125: 9d11882d4b58: Pulling fs layer\n            cluster.go:125: b44d445f805d: Pulling fs layer\n            cluster.go:125: 8bef7a9ebe88: Pulling fs layer\n            cluster.go:125: a0f1c91fa11d: Pulling fs layer\n            cluster.go:125: 2044bb7170df: Pulling fs layer\n            cluster.go:125: 95d77c54a9aa: Pulling fs layer\n            cluster.go:125: c13dd6bf8c10: Pulling fs layer\n            cluster.go:125: a0f1c91fa11d: Waiting\n            cluster.go:125: 2044bb7170df: Waiting\n            cluster.go:125: 95d77c54a9aa: Waiting\n            cluster.go:125: c13dd6bf8c10: Waiting\n            cluster.go:125: 8bef7a9ebe88: Verifying Checksum\n            cluster.go:125: 8bef7a9ebe88: Download complete\n            cluster.go:125: 9d11882d4b58: Verifying Checksum\n            cluster.go:125: 9d11882d4b58: Download complete\n            cluster.go:125: b44d445f805d: Verifying Checksum\n            cluster.go:125: b44d445f805d: Download complete\n            cluster.go:125: a0f1c91fa11d: Download complete\n            cluster.go:125: 2044bb7170df: Download complete\n            cluster.go:125: 95d77c54a9aa: Verifying Checksum\n            cluster.go:125: 95d77c54a9aa: Download complete\n            cluster.go:125: c13dd6bf8c10: Verifying Checksum\n            cluster.go:125: c13dd6bf8c10: Download complete\n            cluster.go:125: 9d11882d4b58: Pull complete\n            cluster.go:125: b44d445f805d: Pull complete\n            cluster.go:125: 8bef7a9ebe88: Pull complete\n            cluster.go:125: a0f1c91fa11d: Pull complete\n            cluster.go:125: 2044bb7170df: Pull complete\n            cluster.go:125: 95d77c54a9aa: Pull complete\n            cluster.go:125: c13dd6bf8c10: Pull complete\n            cluster.go:125: Digest: sha256:6f170eb5f8ea893d94fe3c05cef69270f16614a99d3d922ce8a320efc43b8591\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"linux.nfs.v3","result":"PASS","duration":44362241659,"output":"        nfs.go:80: NFS server booted.\n        nfs.go:85: Test file \"/tmp/tmp.1ZfnpO1VOj\" created on server.\n        nfs.go:125: NFS client booted.\n        nfs.go:133: Got NFS mount.\n"},{"name":"devcontainer.docker","result":"PASS","duration":262021690391,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + ARCH=amd64\n        cluster.go:125: + VERSION=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + EXPECTED_VERSION=4081.3.7\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-2540\n        cluster.go:125: + USR_SRC_DIR=/home/core/dev-container-workdir-2540/src\n        cluster.go:125: + VAR_TMP_DIR=/home/core/dev-container-workdir-2540/tmp\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-2540/src /home/core/dev-container-workdir-2540/tmp\n        cluster.go:125: ++ parted flatcar_developer_container.bin unit b print\n        cluster.go:125: ++ grep Start --after-context=1\n        cluster.go:125: ++ tail --lines=1\n        cluster.go:125: ++ awk '{ print $2 }'\n        cluster.go:125: ++ head --bytes=-2\n        cluster.go:125: + offset=2097152\n        cluster.go:125: + mkdir root\n        cluster.go:125: + sudo mount -o loop,ro,offset=2097152 flatcar_developer_container.bin root\n        cluster.go:125: + sudo tar -C root -czf dev-container-image.tar.gz .\n        cluster.go:125: + sudo umount root\n        cluster.go:125: + rm -f flatcar_developer_container.bin\n        cluster.go:125: + docker import dev-container-image.tar.gz dev-container:42\n        cluster.go:125: + sudo rm -f dev-container-image.tar.gz\n        cluster.go:125: + restorecon /home/core/dev-container-script\n        cluster.go:125: + docker run --log-driver=journald --env PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs --env EXPECTED_VERSION=4081.3.7 --mount type=bind,source=/lib/modules,target=/lib/modules,readonly=true --mount type=bind,source=/home/core/dev-container-script,target=/home/core/dev-container-script,readonly=true --mount type=bind,source=/home/core/dev-container-workdir-2540/src,target=/usr/src --mount type=bind,source=/home/core/dev-container-workdir-2540/tmp,target=/var/tmp dev-container:42 /bin/bash /home/core/dev-container-script\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + [[ 4081.3.7 != \\4\\0\\8\\1\\.\\3\\.\\7 ]]\n        cluster.go:125: + export PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + export 'FEATURES=-ipc-sandbox -network-sandbox'\n        cluster.go:125: + FEATURES='-ipc-sandbox -network-sandbox'\n        cluster.go:125: + emerge-gitclone\n        cluster.go:125: Cloning into '/var/lib/portage/scripts'...\n        cluster.go:125: Note: switching to 'lts-4081.3.7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at a4738dc8a7 New version: lts-4081.3.7\n        cluster.go:125: + emerge --getbinpkg --verbose coreos-sources\n        cluster.go:125: \n        cluster.go:125: !!! Your current profile is deprecated and not supported anymore.\n        cluster.go:125: !!! Use eselect profile to update your profile.\n        cluster.go:125: !!! Please upgrade to the following profile if possible:\n        cluster.go:125: \n        cluster.go:125:         default/linux/amd64/23.0/split-usr/no-multilib/hardened\n        cluster.go:125: \n        cluster.go:125: To upgrade do the following steps:\n        cluster.go:125: \n        cluster.go:125: A profile upgrade to version 23.0 is available for your architecture.\n        cluster.go:125: The new 23.0 profiles enable some toolchain hardening features and \n        cluster.go:125: performance enhancements by default, and standardize settings.\n        cluster.go:125: You can find the list of changes on the wiki tracking page [1].\n        cluster.go:125: \n        cluster.go:125: Upgrade instructions\n        cluster.go:125: \n        cluster.go:125: Note 1: If you have manually changed your CHOST to a value different from \n        cluster.go:125: what the stages and profiles set, you may have to do that in the future too.\n        cluster.go:125: In that case you should know what you are doing, hopefully; please read the \n        cluster.go:125: instructions with a critical eye then.\n        cluster.go:125: \n        cluster.go:125: Note 2: In case you are already familiar with binary packages, you should be\n        cluster.go:125: able to add \"--getbinpkg\" to the emerge calls to speed things up.\n        cluster.go:125: The use of binary packages is completely optional though, and also not\n        cluster.go:125: as much tested as the source-based upgrade path yet.\n        cluster.go:125: \n        cluster.go:125: 1. Ensure your system backups are up to date. Please also update\n        cluster.go:125:    your system fully and depclean before proceeding.\n        cluster.go:125:    glibc older than 2.36 and musl older than 1.2.4 is not supported anymore.\n        cluster.go:125: \n        cluster.go:125: 2. If you are still using one of the long-deprecated amd64 17.0 profiles \n        cluster.go:125:    (other than x32 or musl), then first complete the migration to the \n        cluster.go:125:    corresponding 17.1 profile. Instructions can be found at [3].\n        cluster.go:125:    \n        cluster.go:125: 3. If you are currently using systemd in a split-usr configuration, then first \n        cluster.go:125:    complete the migration to the corresponding merged-usr profile of the \n        cluster.go:125:    same profile version. Details on how to do this can be found in the news \n        cluster.go:125:    item [4].\n        cluster.go:125:    If you are currently using openrc, migrate to 23.0 first, keeping your disk\n        cluster.go:125:    layout. If you want to move from split-usr to merged-usr, do that afterwards.\n        cluster.go:125: \n        cluster.go:125: 4. Run \"emerge --info\" and note down the value of the CHOST variable.\n        cluster.go:125: \n        cluster.go:125: 5. Edit /etc/portage/make.conf; if there is a line defining the CHOST variable,\n        cluster.go:125:    remove it. Also delete all lines defining CHOST_... variables.\n        cluster.go:125: \n        cluster.go:125: 6. Select the 23.0 profile corresponding to your current profile, either using\n        cluster.go:125:    \"eselect profile\" or by manually setting the profile symlink.\n        cluster.go:125:    Note that old profiles are by default split-usr and the 23.0 profiles by\n        cluster.go:125:    default merged-usr. Do NOT change directory scheme now, since this will\n        cluster.go:125:    mess up your system! \n        cluster.go:125:    Instead, make sure that the new profile has the same property: for example, \n        cluster.go:125:    OLD default/linux/amd64/17.1  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/split-usr\n        cluster.go:125:              (added \"split-usr\")\n        cluster.go:125:    OLD default/linux/amd64/17.1/systemd/merged-usr  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/systemd\n        cluster.go:125:              (removed \"merged-usr\")\n        cluster.go:125:    A detailed table of the upgrade paths can be found at [5]. Please consult it.\n        cluster.go:125:    In some cases (hppa, x86) the table will tell you to pick between two choices. \n        cluster.go:125:    What you need should be obvious from your *old* CHOST value (from step 4).\n        cluster.go:125: \n        cluster.go:125: 7. Delete the contents of your binary package cache at ${PKGDIR}\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 8. In the file or directory /etc/portage/binrepos.conf (if existing), update\n        cluster.go:125:    the URI in all configuration such that they point to 23.0 profile binhost \n        cluster.go:125:    directories. The exact paths can be found in the table at [5], too.\n        cluster.go:125: \n        cluster.go:125: 9. Rebuild or reinstall from binary (if available) the following packages in\n        cluster.go:125:    this order, with the same version as already active:\n        cluster.go:125:      emerge --ask --oneshot sys-devel/binutils\n        cluster.go:125:    (you may have to run binutils-config and re-select your binutils now)\n        cluster.go:125:      emerge --ask --oneshot sys-devel/gcc\n        cluster.go:125:    (IMPORTANT: If this command wants to rebuild glibc first, do *not* let it do \n        cluster.go:125:     that; instead, abort and try again with --nodeps added to the command line.)\n        cluster.go:125:    (you may have to run gcc-config and re-select your gcc now)\n        cluster.go:125:    and the C library, i.e. for glibc-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/glibc\n        cluster.go:125:    or for musl-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/musl\n        cluster.go:125: \n        cluster.go:125: 10. Re-run \"emerge --info\" and check if CHOST has changed compared to step 4.\n        cluster.go:125: \n        cluster.go:125: If the CHOST has NOT changed, skip to step 13 (env-update). Otherwise, \n        cluster.go:125: \n        cluster.go:125: 11. Recheck with binutils-config and gcc-config that valid installed versions\n        cluster.go:125:    of binutils and gcc are selected.\n        cluster.go:125: \n        cluster.go:125: 12. Check /etc/env.d, /etc/env.d/binutils, and /etc/env.d/gcc for files that\n        cluster.go:125:    refer to the *OLD* CHOST value, and remove them. \n        cluster.go:125:    Examples how to do this can be found in the similar procedure at [6].\n        cluster.go:125: \n        cluster.go:125: 13. Run env-update \u0026\u0026 source /etc/profile\n        cluster.go:125: \n        cluster.go:125: 14. Re-emerge libtool:\n        cluster.go:125:    emerge --ask --oneshot libtool\n        cluster.go:125: \n        cluster.go:125: 15. Just for safety, delete the contents of your binary package cache at \n        cluster.go:125:     ${PKGDIR} again:\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 16. Rebuild world:\n        cluster.go:125:    emerge --ask --emptytree @world\n        cluster.go:125: \n        cluster.go:125: [1] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_transition\n        cluster.go:125: [2] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_timeline\n        cluster.go:125: [3] https://www.gentoo.org/support/news-items/2019-06-05-amd64-17-1-profiles-are-now-stable.html\n        cluster.go:125: [4] https://www.gentoo.org/support/news-items/2022-12-01-systemd-usrmerge.html\n        cluster.go:125: [5] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_update_table\n        cluster.go:125: [6] https://wiki.gentoo.org/wiki/Changing_the_CHOST_variable#Verifying_things_work\n        cluster.go:125: \n        cluster.go:125: \n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: + zcat /proc/config.gz\n        cluster.go:125: ++ nproc\n        cluster.go:125: + exec make -C /usr/src/linux -j4 modules_prepare V=1\n"},{"name":"cl.disk.raid0.root","result":"PASS","duration":33759932460,"output":""},{"name":"cl.swap_activation","result":"PASS","duration":21253432312,"output":""},{"name":"coreos.selinux.enforce","result":"PASS","duration":43459377778,"output":""},{"name":"cl.locksmith.cluster","result":"PASS","duration":351805578164,"output":""},{"name":"cl.ignition.v1.sethostname","result":"PASS","duration":22435971672,"output":""},{"name":"kubeadm.v1.35.1.cilium.base/node_readiness","result":"PASS","duration":15821428499,"output":""},{"name":"kubeadm.v1.35.1.cilium.base/nginx_deployment","result":"PASS","duration":10801324805,"output":""},{"name":"cl.tpm.root","result":"PASS","duration":53865399863,"output":""},{"name":"coreos.ignition.resource.local","result":"PASS","duration":47894946925,"output":""},{"name":"kubeadm.v1.35.1.cilium.base/NFS_deployment","result":"PASS","duration":11594034761,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"cl.etcd-member.v2-backup-restore","result":"PASS","duration":22325598859,"output":"        cluster.go:125: 2026-04-21T10:25:57Z\tinfo\tetcdutl/backup_command.go:216\tignoring EntryConfChange raft entry\n        cluster.go:125: 2026-04-21T10:25:57Z\tinfo\tetcdutl/backup_command.go:231\tignoring member attribute update on\t{\"entry\": \"Term:2 Index:3 Data:\\\"\\\\010\\\\202\\\\374\\\\342\\\\231\\\\371\\\\265\\\\247\\\\224\\\\033\\\\022\\\\003PUT\\\\032\u0026/0/members/5534734b861a1b28/attributes\\\\\\\"R{\\\\\\\"name\\\\\\\":\\\\\\\"51a696bee63d4765af05adbed2261227\\\\\\\",\\\\\\\"clientURLs\\\\\\\":[\\\\\\\"http://10.0.0.61:2379\\\\\\\"]}(\\\\0002\\\\0008\\\\000H\\\\000P\\\\000X\\\\000`\\\\000h\\\\000p\\\\000x\\\\000\\\\200\\\\001\\\\000\\\" \", \"v2Req.Path\": \"/0/members/5534734b861a1b28/attributes\"}\n        cluster.go:125: 2026-04-21T10:25:57Z\tinfo\tetcdutl/backup_command.go:252\tignoring v3 raft entry\n        cluster.go:125: 2026-04-21T10:25:57Z\tinfo\tmembership/store.go:119\tTrimming membership information from the backend...\n"},{"name":"kubeadm.v1.35.1.cilium.base/IPSec_encryption","result":"PASS","duration":17145245193,"output":"        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"kubeadm.v1.35.1.cilium.base","result":"PASS","duration":137455060469,"output":"        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.13.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.6-0\n        cluster.go:125: [init] Using Kubernetes version: v1.35.4\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:24:40.537220    2016 checks.go:906] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.55]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 501.447014ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.55:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.506466945s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 2.049016178s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 3.501785941s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 8qcrff.veoc7voji21cxdyl\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.55:6443 --token 8qcrff.veoc7voji21cxdyl \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:df071b116d334edeea9cc40e8b0f7d30f62739a5b088631ded7366622c9b3187 \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0421 10:25:09.467006    1660 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.35.1.cilium.base/node_readiness (15.82s)\n    --- PASS: kubeadm.v1.35.1.cilium.base/nginx_deployment (10.80s)\n    --- PASS: kubeadm.v1.35.1.cilium.base/NFS_deployment (11.59s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n    --- PASS: kubeadm.v1.35.1.cilium.base/IPSec_encryption (17.15s)\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"systemd.sysusers.gshadow","result":"PASS","duration":22713408493,"output":""},{"name":"cl.ignition.oem.regular","result":"PASS","duration":44663422668,"output":""},{"name":"docker.userns","result":"PASS","duration":24484740406,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.24MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:82c416df15076852d26cc96651eff5f8bca2c0bf4e83d8c92f18cc74d85da693 done\n        cluster.go:125: #5 naming to docker.io/library/userns-test done\n        cluster.go:125: #5 DONE 0.0s\n"},{"name":"cl.filesystem/deadlinks","result":"PASS","duration":1178358405,"output":""},{"name":"cl.filesystem/suid","result":"PASS","duration":186498192,"output":""},{"name":"cl.filesystem/sgid","result":"PASS","duration":181871524,"output":""},{"name":"cl.filesystem/writablefiles","result":"PASS","duration":176825929,"output":""},{"name":"cl.filesystem/writabledirs","result":"PASS","duration":145077444,"output":""},{"name":"cl.filesystem/stickydirs","result":"PASS","duration":167868339,"output":""},{"name":"cl.filesystem/blacklist","result":"PASS","duration":181929209,"output":""},{"name":"cl.filesystem","result":"PASS","duration":24389772016,"output":"    --- PASS: cl.filesystem/deadlinks (1.18s)\n    --- PASS: cl.filesystem/suid (0.19s)\n    --- PASS: cl.filesystem/sgid (0.18s)\n    --- PASS: cl.filesystem/writablefiles (0.18s)\n    --- PASS: cl.filesystem/writabledirs (0.15s)\n    --- PASS: cl.filesystem/stickydirs (0.17s)\n    --- PASS: cl.filesystem/blacklist (0.18s)\n"},{"name":"docker.network-openbsd-nc","result":"PASS","duration":26891042744,"output":"        docker.go:413: creating netcat containers\n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.45MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:55af17be53e9be31698bd2a955507bdfca3f152b7456a1a7357f3a8fce8977fe done\n        cluster.go:125: #5 naming to docker.io/library/netcat done\n        cluster.go:125: #5 DONE 0.0s\n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.45MB 0.0s done\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:3063a4f2c674a4323b82d361f0afcc46e732b9277ca01ca5e38fc51ed8c5262b done\n        cluster.go:125: #5 naming to docker.io/library/netcat done\n        cluster.go:125: #5 DONE 0.0s\n"},{"name":"packages/sys-cluster/ipvsadm","result":"PASS","duration":570175981,"output":""},{"name":"cl.install.cloudinit","result":"PASS","duration":23491087219,"output":""},{"name":"cl.tang.nonroot","result":"PASS","duration":62687088428,"output":"        tang.go:148: Started tang on 10.0.0.1:44735\n"},{"name":"packages/sys-block/open-iscsi","result":"FAIL","duration":29958018803,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/targetcli-fb:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/targetcli-fb\n        cluster.go:125: a2318d6c47ec: Pulling fs layer\n        cluster.go:125: 3d3086a1439f: Pulling fs layer\n        cluster.go:125: 3d3086a1439f: Download complete\n        cluster.go:125: a2318d6c47ec: Verifying Checksum\n        cluster.go:125: a2318d6c47ec: Download complete\n        cluster.go:125: a2318d6c47ec: Pull complete\n        cluster.go:125: 3d3086a1439f: Pull complete\n        cluster.go:125: Digest: sha256:b6cd65db981974e8b74938617218dd023775b969f9a059ced21e6ce6fa4763c1\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/targetcli-fb:latest\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: Created symlink /etc/systemd/system/remote-fs.target.wants/iscsi.service → /usr/lib/systemd/system/iscsi.service.\n        cluster.go:145: \"sudo /check\" failed: output no /dev/sda device after reboot, status Process exited with status 1\n"},{"name":"packages","result":"FAIL","duration":53899019525,"output":"    --- PASS: packages/sys-cluster/ipvsadm (0.57s)\n    --- FAIL: packages/sys-block/open-iscsi (29.96s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/targetcli-fb:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/targetcli-fb\n            cluster.go:125: a2318d6c47ec: Pulling fs layer\n            cluster.go:125: 3d3086a1439f: Pulling fs layer\n            cluster.go:125: 3d3086a1439f: Download complete\n            cluster.go:125: a2318d6c47ec: Verifying Checksum\n            cluster.go:125: a2318d6c47ec: Download complete\n            cluster.go:125: a2318d6c47ec: Pull complete\n            cluster.go:125: 3d3086a1439f: Pull complete\n            cluster.go:125: Digest: sha256:b6cd65db981974e8b74938617218dd023775b969f9a059ced21e6ce6fa4763c1\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/targetcli-fb:latest\n            cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n            cluster.go:125: Created symlink /etc/systemd/system/remote-fs.target.wants/iscsi.service → /usr/lib/systemd/system/iscsi.service.\n            cluster.go:145: \"sudo /check\" failed: output no /dev/sda device after reboot, status Process exited with status 1\n"},{"name":"coreos.selinux.boolean","result":"PASS","duration":45991393036,"output":""},{"name":"coreos.locksmith.tls","result":"PASS","duration":45556726290,"output":""},{"name":"linux.nfs.v4","result":"PASS","duration":42736667105,"output":"        nfs.go:80: NFS server booted.\n        nfs.go:85: Test file \"/tmp/tmp.G4h4dd9xjM\" created on server.\n        nfs.go:125: NFS client booted.\n        nfs.go:133: Got NFS mount.\n"},{"name":"cl.ignition.v2_1.vfat","result":"PASS","duration":22428633681,"output":""},{"name":"misc.fips","result":"PASS","duration":34090022193,"output":"        cluster.go:125: Error setting digest\n        cluster.go:125: 40879015257F0000:error:0308010C:digital envelope routines:inner_evp_generic_fetch:unsupported:../openssl-3.2.3/crypto/evp/evp_fetch.c:355:Global default library context, Algorithm (MD5 : 102), Properties ()\n        cluster.go:125: 40879015257F0000:error:03000086:digital envelope routines:evp_md_init_internal:initialization error:../openssl-3.2.3/crypto/evp/digest.c:272:\n        cluster.go:152: + cat /proc/sys/crypto/fips_enabled\n"},{"name":"docker.containerd-restart","result":"PASS","duration":77858169709,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Download complete\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n"},{"name":"cl.ignition.luks","result":"PASS","duration":31457527611,"output":""},{"name":"kubeadm.v1.33.8.calico.cgroupv1.base/node_readiness","result":"PASS","duration":26365230438,"output":""},{"name":"kubeadm.v1.33.8.calico.cgroupv1.base/nginx_deployment","result":"PASS","duration":5508882286,"output":""},{"name":"kubeadm.v1.33.8.calico.cgroupv1.base/NFS_deployment","result":"PASS","duration":11682617021,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.33.8.calico.cgroupv1.base","result":"PASS","duration":143834790986,"output":"        cluster.go:125: I0421 10:30:31.620977    1945 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0421 10:30:38.130904    2169 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:30:38.222035    2169 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.82]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 500.850819ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.82:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.797004528s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 3.159460912s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 5.003621181s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: d4gjdx.4j6jyyhksoca6dms\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.82:6443 --token d4gjdx.4j6jyyhksoca6dms \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:0c1d7c68b6ae106b0bb3c893ca51b3251643d3c07a2b9aa72bcac12bf8627737 \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: installation.operator.tigera.io/default created\n        cluster.go:125: apiserver.operator.tigera.io/default created\n        cluster.go:125: goldmane.operator.tigera.io/default created\n        cluster.go:125: whisker.operator.tigera.io/default created\n        cluster.go:125: W0421 10:31:24.529597    1787 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.calico.cgroupv1.base/node_readiness (26.37s)\n    --- PASS: kubeadm.v1.33.8.calico.cgroupv1.base/nginx_deployment (5.51s)\n    --- PASS: kubeadm.v1.33.8.calico.cgroupv1.base/NFS_deployment (11.68s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"cl.update.badverity","result":"PASS","duration":326142891623,"output":""},{"name":"cl.tpm.root-cryptenroll-pcr-noupdate","result":"PASS","duration":152390644799,"output":""},{"name":"cl.ignition.v2.ext4root","result":"PASS","duration":21589306035,"output":""},{"name":"linux.ntp","result":"PASS","duration":21406605049,"output":""},{"name":"cl.internet/UpdateEngine","result":"PASS","duration":134350852,"output":""},{"name":"cl.internet/DockerPing","result":"PASS","duration":5988880797,"output":""},{"name":"cl.internet/DockerEcho","result":"PASS","duration":436999108,"output":""},{"name":"cl.internet/NTPDate","result":"PASS","duration":6856050190,"output":""},{"name":"cl.internet","result":"PASS","duration":34995983027,"output":"    --- PASS: cl.internet/UpdateEngine (0.13s)\n    --- PASS: cl.internet/DockerPing (5.99s)\n    --- PASS: cl.internet/DockerEcho (0.44s)\n    --- PASS: cl.internet/NTPDate (6.86s)\n"},{"name":"cl.flannel.udp","result":"PASS","duration":116699861236,"output":"        cluster.go:125: Timeout occurred while waiting for network connectivity.\n        flannel.go:121: ping from 2b9f8b1b-7c8a-45ce-a962-4e722e83fdc7(10.254.58.0) to 4c49c990-b49d-40c9-bbcd-684c82e19c11(10.254.15.1)\n"},{"name":"sysext.simple","result":"PASS","duration":22399234472,"output":""},{"name":"kubeadm.v1.33.8.cilium.cgroupv1.base/node_readiness","result":"PASS","duration":10652449433,"output":""},{"name":"coreos.auth.verify","result":"PASS","duration":22687759183,"output":""},{"name":"kubeadm.v1.33.8.cilium.cgroupv1.base/nginx_deployment","result":"PASS","duration":15841858621,"output":""},{"name":"kubeadm.v1.33.8.cilium.cgroupv1.base/NFS_deployment","result":"PASS","duration":16635032774,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"sysext.disable-docker","result":"PASS","duration":24986753706,"output":""},{"name":"kubeadm.v1.33.8.cilium.cgroupv1.base/IPSec_encryption","result":"PASS","duration":18919495730,"output":"        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"kubeadm.v1.33.8.cilium.cgroupv1.base","result":"PASS","duration":142872871859,"output":"        cluster.go:125: I0421 10:33:53.317777    1940 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0421 10:33:59.647961    2163 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:33:59.742741    2163 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.92]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 501.707752ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.92:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.504389679s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 2.907405687s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 4.503212617s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: a5vvsj.hbo07g3irqmzd14s\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.92:6443 --token a5vvsj.hbo07g3irqmzd14s \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:b3fda1dba1a88f5a779c0e350b7cf383eb01c8353289446a7bbdf455f2f85317 \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0421 10:34:30.374599    1788 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING SystemVerification]: cgroups v1 support is in maintenance mode, please migrate to cgroups v2\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.cilium.cgroupv1.base/node_readiness (10.65s)\n    --- PASS: kubeadm.v1.33.8.cilium.cgroupv1.base/nginx_deployment (15.84s)\n    --- PASS: kubeadm.v1.33.8.cilium.cgroupv1.base/NFS_deployment (16.64s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n    --- PASS: kubeadm.v1.33.8.cilium.cgroupv1.base/IPSec_encryption (18.92s)\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"kubeadm.v1.34.4.cilium.base/node_readiness","result":"PASS","duration":10628361492,"output":""},{"name":"cl.etcd-member.etcdctlv3","result":"PASS","duration":28101612713,"output":"        cluster.go:125: {\"level\":\"info\",\"ts\":1776767745.267989,\"caller\":\"snapshot/v3_snapshot.go:68\",\"msg\":\"created temporary db file\",\"path\":\"/tmp/tmp.aiy8LlAEuH/snapshot.db.part\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776767745.2686582,\"logger\":\"client\",\"caller\":\"v3/maintenance.go:211\",\"msg\":\"opened snapshot stream; downloading\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776767745.2686837,\"caller\":\"snapshot/v3_snapshot.go:76\",\"msg\":\"fetching snapshot\",\"endpoint\":\"127.0.0.1:2379\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776767745.2718248,\"logger\":\"client\",\"caller\":\"v3/maintenance.go:219\",\"msg\":\"completed snapshot read; closing\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776767745.2719367,\"caller\":\"snapshot/v3_snapshot.go:91\",\"msg\":\"fetched snapshot\",\"endpoint\":\"127.0.0.1:2379\",\"size\":\"20 kB\",\"took\":\"now\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776767745.2720115,\"caller\":\"snapshot/v3_snapshot.go:100\",\"msg\":\"saved\",\"path\":\"/tmp/tmp.aiy8LlAEuH/snapshot.db\"}\n        cluster.go:125: Deprecated: Use `etcdutl snapshot status` instead.\n"},{"name":"kubeadm.v1.34.4.cilium.base/nginx_deployment","result":"PASS","duration":10782186664,"output":""},{"name":"kubeadm.v1.34.4.cilium.base/NFS_deployment","result":"PASS","duration":17061590554,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"cl.network.initramfs.second-boot","result":"PASS","duration":42617111372,"output":""},{"name":"kubeadm.v1.34.4.cilium.base/IPSec_encryption","result":"PASS","duration":18884075172,"output":"        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"kubeadm.v1.34.4.cilium.base","result":"PASS","duration":161542585408,"output":"        cluster.go:125: I0421 10:34:47.348038    1799 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0421 10:34:55.661238    2018 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:34:55.754144    2018 checks.go:827] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.96]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.001602443s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.96:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 2.053748303s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 3.524947648s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 5.002947916s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 2xyz68.d84lci4fylaoi2i0\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.96:6443 --token 2xyz68.d84lci4fylaoi2i0 \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:6ca3a53a8ad1ac2732577b80d434b7a12185b9e8f23d934eac510cdf7d65ff8a \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0421 10:35:33.826762    1657 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.34.4.cilium.base/node_readiness (10.63s)\n    --- PASS: kubeadm.v1.34.4.cilium.base/nginx_deployment (10.78s)\n    --- PASS: kubeadm.v1.34.4.cilium.base/NFS_deployment (17.06s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n    --- PASS: kubeadm.v1.34.4.cilium.base/IPSec_encryption (18.88s)\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"sysext.custom-oem","result":"PASS","duration":63347650292,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-25255\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-25255\n        cluster.go:125: + sudo systemd-nspawn --console=pipe --bind-ro=/home/core/dev-container-script.sh --bind=/home/core/dev-container-workdir-25255:/work --image=flatcar_developer_container.bin --machine=flatcar-developer-container /bin/bash /home/core/dev-container-script.sh\n        cluster.go:125: + set -euo pipefail\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-21-0856\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-21-0856\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version_id=4081.3.7\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64-usr\n        cluster.go:125: + board=amd64-usr\n        cluster.go:125: + mkdir -p /work/sysext_rootfs/usr/share/flatcar-sysext-kola-test\n        cluster.go:125: + echo 4081.3.7\n        cluster.go:125: + mkdir -p /work/sysext_rootfs/usr/lib/extension-release.d\n        cluster.go:125: + sysext_arch=x86-64\n        cluster.go:125: + [[ amd64-usr = \\a\\r\\m\\6\\4\\-\\u\\s\\r ]]\n        cluster.go:125: + metadata=('ID=flatcar' \"VERSION_ID=${version_id}\" \"ARCHITECTURE=${sysext_arch}\")\n        cluster.go:125: + metadata_file=/work/sysext_rootfs/usr/lib/extension-release.d/extension-release.oem-test\n        cluster.go:125: + printf '%s\\n' ID=flatcar VERSION_ID=4081.3.7 ARCHITECTURE=x86-64\n        cluster.go:125: + mksquashfs /work/sysext_rootfs /work/oem-test-4081.3.7.raw -all-root\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-21-0909\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: + sysext_file=/home/core/dev-container-workdir-25255/oem-test-4081.3.7.raw\n        cluster.go:125: + [[ ! -e /home/core/dev-container-workdir-25255/oem-test-4081.3.7.raw ]]\n        cluster.go:125: + [[ ! -e /oem/oem-release ]]\n        cluster.go:125: + printf '%s\\n' ID=test VERSION_ID=1.0.0 'NAME=testing stuff'\n        cluster.go:125: + sudo tee /oem/oem-release\n        cluster.go:125: + sudo mkdir -p /oem/sysext\n        cluster.go:125: + sudo mv /home/core/dev-container-workdir-25255/oem-test-4081.3.7.raw /oem/sysext\n        cluster.go:125: + sudo touch /oem/sysext/active-oem-test\n        cluster.go:125: + sudo flatcar-reset --keep-machine-id --keep-paths /var/log\n        cluster.go:125: + set -euo pipefail\n        cluster.go:125: ++ systemd-sysext list --json=pretty\n        cluster.go:125: + list_out='[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"containerd-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/containerd-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776763029925530\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"docker-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/docker-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776763058714418\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"oem-test\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/oem-test.raw\",\n        cluster.go:125: \t\t\"time\" : 1776767785821723\n        cluster.go:125: \t}\n        cluster.go:125: ]'\n        cluster.go:125: ++ systemd-sysext status --json=pretty\n        cluster.go:125: + status_out='[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/opt\",\n        cluster.go:125: \t\t\"extensions\" : \"none\",\n        cluster.go:125: \t\t\"since\" : null\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/usr\",\n        cluster.go:125: \t\t\"extensions\" : [\n        cluster.go:125: \t\t\t\"containerd-flatcar\",\n        cluster.go:125: \t\t\t\"docker-flatcar\",\n        cluster.go:125: \t\t\t\"oem-test\"\n        cluster.go:125: \t\t],\n        cluster.go:125: \t\t\"since\" : 1776767796432000\n        cluster.go:125: \t}\n        cluster.go:125: ]'\n        cluster.go:125: + printf 'sysext list:\\n%s\\nsysext status:\\n%s\\n' '[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"containerd-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/containerd-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776763029925530\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"docker-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/docker-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776763058714418\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"oem-test\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/oem-test.raw\",\n        cluster.go:125: \t\t\"time\" : 1776767785821723\n        cluster.go:125: \t}\n        cluster.go:125: ]' '[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/opt\",\n        cluster.go:125: \t\t\"extensions\" : \"none\",\n        cluster.go:125: \t\t\"since\" : null\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/usr\",\n        cluster.go:125: \t\t\"extensions\" : [\n        cluster.go:125: \t\t\t\"containerd-flatcar\",\n        cluster.go:125: \t\t\t\"docker-flatcar\",\n        cluster.go:125: \t\t\t\"oem-test\"\n        cluster.go:125: \t\t],\n        cluster.go:125: \t\t\"since\" : 1776767796432000\n        cluster.go:125: \t}\n        cluster.go:125: ]'\n        cluster.go:125: ++ jq '.[] | select(.name == \"oem-test\")'\n        cluster.go:125: + list_oem_test='{\n        cluster.go:125:   \"name\": \"oem-test\",\n        cluster.go:125:   \"type\": \"raw\",\n        cluster.go:125:   \"path\": \"/etc/extensions/oem-test.raw\",\n        cluster.go:125:   \"time\": 1776767785821723\n        cluster.go:125: }'\n        cluster.go:125: + [[ -z {\n        cluster.go:125:   \"name\": \"oem-test\",\n        cluster.go:125:   \"type\": \"raw\",\n        cluster.go:125:   \"path\": \"/etc/extensions/oem-test.raw\",\n        cluster.go:125:   \"time\": 1776767785821723\n        cluster.go:125: } ]]\n        cluster.go:125: ++ jq --raw-output .type\n        cluster.go:125: + oem_test_type=raw\n        cluster.go:125: + [[ raw != \\r\\a\\w ]]\n        cluster.go:125: ++ jq --raw-output .path\n        cluster.go:125: + oem_test_path=/etc/extensions/oem-test.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + oem_test_real_path=/oem/sysext/oem-test-4081.3.7.raw\n        cluster.go:125: + [[ /etc/extensions/oem-test.raw != \\/\\o\\e\\m\\/\\s\\y\\s\\e\\x\\t\\/\\o\\e\\m\\-\\t\\e\\s\\t\\-\\4\\0\\8\\1\\.\\3\\.\\7\\.\\r\\a\\w ]]\n        cluster.go:125: + [[ /etc/extensions/oem-test.raw != \\/\\e\\t\\c\\/\\e\\x\\t\\e\\n\\s\\i\\o\\n\\s\\/\\o\\e\\m\\-\\t\\e\\s\\t\\.\\r\\a\\w ]]\n        cluster.go:125: ++ jq '.[] | select(.hierarchy == \"/usr\")'\n        cluster.go:125: + status_usr='{\n        cluster.go:125:   \"hierarchy\": \"/usr\",\n        cluster.go:125:   \"extensions\": [\n        cluster.go:125:     \"containerd-flatcar\",\n        cluster.go:125:     \"docker-flatcar\",\n        cluster.go:125:     \"oem-test\"\n        cluster.go:125:   ],\n        cluster.go:125:   \"since\": 1776767796432000\n        cluster.go:125: }'\n        cluster.go:125: + [[ -z {\n        cluster.go:125:   \"hierarchy\": \"/usr\",\n        cluster.go:125:   \"extensions\": [\n        cluster.go:125:     \"containerd-flatcar\",\n        cluster.go:125:     \"docker-flatcar\",\n        cluster.go:125:     \"oem-test\"\n        cluster.go:125:   ],\n        cluster.go:125:   \"since\": 1776767796432000\n        cluster.go:125: } ]]\n        cluster.go:125: ++ jq --raw-output '.extensions[] | select(. == \"oem-test\")'\n        cluster.go:125: + status_usr_extensions_oem_test=oem-test\n        cluster.go:125: + [[ oem-test != \\o\\e\\m\\-\\t\\e\\s\\t ]]\n        cluster.go:125: + f=/usr/share/flatcar-sysext-kola-test/file\n        cluster.go:125: + [[ ! -e /usr/share/flatcar-sysext-kola-test/file ]]\n        cluster.go:125: ++ cat /usr/share/flatcar-sysext-kola-test/file\n        cluster.go:125: + got=4081.3.7\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-21-0909\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + ex=4081.3.7\n        cluster.go:125: + [[ 4081.3.7 != \\4\\0\\8\\1\\.\\3\\.\\7 ]]\n"},{"name":"cl.ignition.v2.btrfsroot","result":"PASS","duration":22548860811,"output":""},{"name":"cl.misc.falco","result":"PASS","duration":48053709656,"output":"        cluster.go:125: Unable to find image 'falcosecurity/falco-driver-loader:master' locally\n        cluster.go:125: master: Pulling from falcosecurity/falco-driver-loader\n        cluster.go:125: 0858e4838ecf: Pulling fs layer\n        cluster.go:125: 80a4111ac8b3: Pulling fs layer\n        cluster.go:125: 7ca02b979ec3: Pulling fs layer\n        cluster.go:125: 063817c6b581: Pulling fs layer\n        cluster.go:125: 3cf8fec6def5: Pulling fs layer\n        cluster.go:125: 422fe6f8a0ed: Pulling fs layer\n        cluster.go:125: fde222b00d03: Pulling fs layer\n        cluster.go:125: 08cd6f200908: Pulling fs layer\n        cluster.go:125: 422fe6f8a0ed: Waiting\n        cluster.go:125: 063817c6b581: Waiting\n        cluster.go:125: 3cf8fec6def5: Waiting\n        cluster.go:125: fde222b00d03: Waiting\n        cluster.go:125: 08cd6f200908: Waiting\n        cluster.go:125: 0858e4838ecf: Verifying Checksum\n        cluster.go:125: 0858e4838ecf: Download complete\n        cluster.go:125: 7ca02b979ec3: Verifying Checksum\n        cluster.go:125: 7ca02b979ec3: Download complete\n        cluster.go:125: 80a4111ac8b3: Verifying Checksum\n        cluster.go:125: 80a4111ac8b3: Download complete\n        cluster.go:125: 063817c6b581: Verifying Checksum\n        cluster.go:125: 063817c6b581: Download complete\n        cluster.go:125: 3cf8fec6def5: Verifying Checksum\n        cluster.go:125: 3cf8fec6def5: Download complete\n        cluster.go:125: fde222b00d03: Verifying Checksum\n        cluster.go:125: fde222b00d03: Download complete\n        cluster.go:125: 08cd6f200908: Verifying Checksum\n        cluster.go:125: 08cd6f200908: Download complete\n        cluster.go:125: 0858e4838ecf: Pull complete\n        cluster.go:125: 80a4111ac8b3: Pull complete\n        cluster.go:125: 7ca02b979ec3: Pull complete\n        cluster.go:125: 063817c6b581: Pull complete\n        cluster.go:125: 3cf8fec6def5: Pull complete\n        cluster.go:125: 422fe6f8a0ed: Verifying Checksum\n        cluster.go:125: 422fe6f8a0ed: Download complete\n        cluster.go:125: 422fe6f8a0ed: Pull complete\n        cluster.go:125: fde222b00d03: Pull complete\n        cluster.go:125: 08cd6f200908: Pull complete\n        cluster.go:125: Digest: sha256:c947fdaf0be8c9c20e530f87d551d9303783be950776b6f2ced50f7f0672e1a1\n        cluster.go:125: Status: Downloaded newer image for falcosecurity/falco-driver-loader:master\n"},{"name":"cl.cgroupv1/CgroupMounts","result":"PASS","duration":122501401,"output":""},{"name":"cl.cgroupv1","result":"PASS","duration":22599637403,"output":"    --- PASS: cl.cgroupv1/CgroupMounts (0.12s)\n"},{"name":"docker.btrfs-storage","result":"PASS","duration":22860678800,"output":""},{"name":"cl.ignition.translation","result":"PASS","duration":23570217027,"output":"        cluster.go:152: + ip --json address show kola | jq -r '.[] | .addr_info | .[] | select( .family == \"inet\") | .local'\n        cluster.go:152: + cat /etc/systemd/network/00-dummy.network\n"},{"name":"docker.enable-service.sysext","result":"PASS","duration":22326231985,"output":"        cluster.go:152: + systemctl is-enabled docker\n"},{"name":"sysext.custom-docker.sysext","result":"PASS","duration":36556846854,"output":"        cluster.go:125: bash: line 1: docker: command not found\n        cluster.go:125: Cloning into 'sysext-bakery'...\n        cluster.go:125: Note: switching to '9850ffd5b2353f45a9b3bf4fb84f8138a149e3e7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at 9850ffd Merge pull request #31 from flatcar/t-lo/fix-docker-23-containerd-shim\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Download complete\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n"},{"name":"cl.cloudinit.basic","result":"PASS","duration":21438633827,"output":""},{"name":"coreos.ignition.resource.remote","result":"PASS","duration":22531884416,"output":""},{"name":"coreos.ignition.resource.s3.versioned","result":"PASS","duration":22415595881,"output":""},{"name":"cl.ignition.oem.reuse","result":"PASS","duration":43552816444,"output":""},{"name":"cl.tpm.root-cryptenroll","result":"PASS","duration":53144962926,"output":""},{"name":"kubeadm.v1.35.1.calico.base/node_readiness","result":"PASS","duration":15867822681,"output":""},{"name":"kubeadm.v1.35.1.calico.base/nginx_deployment","result":"PASS","duration":5694614822,"output":""},{"name":"systemd.journal.remote","result":"PASS","duration":47229985121,"output":"        cluster.go:125: Running as unit: systemd-journal-remote-client.service; invocation ID: 4d9e9849569e4c5f88782611b81113d7\n"},{"name":"kubeadm.v1.35.1.calico.base/NFS_deployment","result":"PASS","duration":17265551110,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.35.1.calico.base","result":"PASS","duration":131496971869,"output":"        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.13.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.6-0\n        cluster.go:125: [init] Using Kubernetes version: v1.35.4\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:38:30.589737    2013 checks.go:906] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.114]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 501.00317ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.114:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.505287741s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 1.588080049s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 3.001700555s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: gtqgs0.gxke2a1q0hitahrb\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.114:6443 --token gtqgs0.gxke2a1q0hitahrb \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:306cc16f7777cbb2a084d199633642118ed5f19ba1aaa1ff849c31d506348a16 \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: installation.operator.tigera.io/default created\n        cluster.go:125: apiserver.operator.tigera.io/default created\n        cluster.go:125: goldmane.operator.tigera.io/default created\n        cluster.go:125: whisker.operator.tigera.io/default created\n        cluster.go:125: W0421 10:39:12.041296    1646 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.35.1.calico.base/node_readiness (15.87s)\n    --- PASS: kubeadm.v1.35.1.calico.base/nginx_deployment (5.69s)\n    --- PASS: kubeadm.v1.35.1.calico.base/NFS_deployment (17.27s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"cl.network.listeners","result":"PASS","duration":22360303686,"output":""},{"name":"cl.flannel.vxlan","result":"PASS","duration":50689857242,"output":"        flannel.go:121: ping from 2e8f07d7-dfc5-41e9-b442-a68213085a27(10.254.97.0) to 824960d4-1e87-4983-9d77-0a3961039297(10.254.34.0)\n"},{"name":"cl.ignition.v2.xfsroot","result":"PASS","duration":22605003637,"output":""},{"name":"cl.ignition.instantiated.enable-unit","result":"PASS","duration":21831463280,"output":""},{"name":"cl.ignition.v1.xfsroot","result":"PASS","duration":22855429682,"output":""},{"name":"cl.ignition.v2_1.ext4checkexisting","result":"PASS","duration":45679176935,"output":""},{"name":"kubeadm.v1.33.8.cilium.base/node_readiness","result":"PASS","duration":10539646439,"output":""},{"name":"kubeadm.v1.33.8.cilium.base/nginx_deployment","result":"PASS","duration":10814869643,"output":""},{"name":"docker.selinux","result":"PASS","duration":25725438590,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Download complete\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: sh: can't create /opt/hello: Permission denied\n"},{"name":"kubeadm.v1.33.8.cilium.base/NFS_deployment","result":"PASS","duration":11962859071,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"cl.basic/Script","result":"PASS","duration":214334834,"output":""},{"name":"cl.basic/PortSSH","result":"PASS","duration":128010209,"output":""},{"name":"cl.basic/DbusPerms","result":"PASS","duration":155387632,"output":""},{"name":"cl.basic/Symlink","result":"PASS","duration":135178653,"output":""},{"name":"cl.basic/SymlinkFlatcar","result":"PASS","duration":126121656,"output":""},{"name":"cl.basic/ReadOnly","result":"PASS","duration":127936685,"output":""},{"name":"cl.basic/MachineID","result":"PASS","duration":136923859,"output":""},{"name":"cl.basic/UpdateEngineKeys","result":"PASS","duration":134442543,"output":""},{"name":"cl.basic/ServicesActive","result":"PASS","duration":137102204,"output":""},{"name":"cl.basic/RandomUUID","result":"PASS","duration":155417655,"output":""},{"name":"cl.basic/Useradd","result":"PASS","duration":185819520,"output":""},{"name":"cl.basic/Microcode","result":"PASS","duration":139999826,"output":""},{"name":"cl.basic/CloudConfig","result":"PASS","duration":356927095,"output":""},{"name":"cl.basic","result":"PASS","duration":23817841405,"output":"    --- PASS: cl.basic/Script (0.21s)\n    --- PASS: cl.basic/PortSSH (0.13s)\n    --- PASS: cl.basic/DbusPerms (0.16s)\n    --- PASS: cl.basic/Symlink (0.14s)\n    --- PASS: cl.basic/SymlinkFlatcar (0.13s)\n    --- PASS: cl.basic/ReadOnly (0.13s)\n    --- PASS: cl.basic/MachineID (0.14s)\n    --- PASS: cl.basic/UpdateEngineKeys (0.13s)\n    --- PASS: cl.basic/ServicesActive (0.14s)\n    --- PASS: cl.basic/RandomUUID (0.16s)\n    --- PASS: cl.basic/Useradd (0.19s)\n    --- PASS: cl.basic/Microcode (0.14s)\n    --- PASS: cl.basic/CloudConfig (0.36s)\n"},{"name":"kubeadm.v1.33.8.cilium.base/IPSec_encryption","result":"PASS","duration":19676228447,"output":"        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"kubeadm.v1.33.8.cilium.base","result":"PASS","duration":137230559524,"output":"        cluster.go:125: I0421 10:41:25.442574    1798 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0421 10:41:31.314827    2017 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:41:31.409234    2017 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.129]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 503.048588ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.129:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.50491967s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 1.7299459s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 3.50202513s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: tzicyg.7ojjr021al5xlyvz\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.129:6443 --token tzicyg.7ojjr021al5xlyvz \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:e64e8d8968f45234f9311f6eda4e5cfd6886d5aa22eff92d4463ea941e44d83a \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0421 10:42:00.047822    1649 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.cilium.base/node_readiness (10.54s)\n    --- PASS: kubeadm.v1.33.8.cilium.base/nginx_deployment (10.81s)\n    --- PASS: kubeadm.v1.33.8.cilium.base/NFS_deployment (11.96s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n    --- PASS: kubeadm.v1.33.8.cilium.base/IPSec_encryption (19.68s)\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"cl.ignition.oem.wipe","result":"PASS","duration":22449876262,"output":""},{"name":"cl.sysext.fallbackdownload","result":"SKIP","duration":22637064629,"output":"        update.go:652: Test skipped, only dev builds are supported\n"},{"name":"cl.update.reboot","result":"PASS","duration":37190577856,"output":""},{"name":"cl.ignition.v2_1.swap","result":"PASS","duration":21458472014,"output":""},{"name":"coreos.ignition.groups","result":"PASS","duration":21636687725,"output":""},{"name":"cl.cloudinit.script","result":"PASS","duration":21373782667,"output":""},{"name":"cl.etcd-member.discovery","result":"PASS","duration":35148527358,"output":""},{"name":"coreos.update.badusr","result":"PASS","duration":347731679983,"output":""},{"name":"coreos.tls.fetch-urls","result":"PASS","duration":25169696848,"output":""},{"name":"kubeadm.v1.33.8.flannel.base/node_readiness","result":"PASS","duration":10492819719,"output":""},{"name":"kubeadm.v1.33.8.flannel.base/nginx_deployment","result":"PASS","duration":5682347473,"output":""},{"name":"coreos.ignition.once","result":"PASS","duration":44281541798,"output":""},{"name":"sysext.disable-containerd","result":"PASS","duration":23429610607,"output":""},{"name":"kubeadm.v1.33.8.flannel.base/NFS_deployment","result":"PASS","duration":17561775869,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.33.8.flannel.base","result":"PASS","duration":116494327243,"output":"        cluster.go:125: I0421 10:44:37.484862    1762 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0421 10:44:46.017006    1982 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:44:46.107331    1982 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.142]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 501.768331ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.142:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.562528848s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 1.845132212s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 3.502114348s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: m34hls.d0qdmf8owt2swqkh\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.142:6443 --token m34hls.d0qdmf8owt2swqkh \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:c5ca78ad4753ee150a9f85a95f114a83a27ce02235f4adb806af0ed9d2bcdff6 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0421 10:45:15.084081    1604 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.flannel.base/node_readiness (10.49s)\n    --- PASS: kubeadm.v1.33.8.flannel.base/nginx_deployment (5.68s)\n    --- PASS: kubeadm.v1.33.8.flannel.base/NFS_deployment (17.56s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"coreos.ignition.systemd.enable-service","result":"PASS","duration":21893994614,"output":""},{"name":"cl.disk.raid1.data","result":"PASS","duration":45960415037,"output":""},{"name":"cl.ignition.v1.once","result":"PASS","duration":47203426413,"output":""},{"name":"cl.tpm.nonroot","result":"PASS","duration":63391538929,"output":""},{"name":"cl.ignition.v1.groups","result":"PASS","duration":21608784707,"output":""},{"name":"kubeadm.v1.35.1.flannel.base/node_readiness","result":"PASS","duration":15947933806,"output":""},{"name":"kubeadm.v1.33.8.calico.base/node_readiness","result":"PASS","duration":32255519885,"output":""},{"name":"kubeadm.v1.35.1.flannel.base/nginx_deployment","result":"PASS","duration":10988473797,"output":""},{"name":"kubeadm.v1.33.8.calico.base/nginx_deployment","result":"PASS","duration":5834358147,"output":""},{"name":"kubeadm.v1.35.1.flannel.base/NFS_deployment","result":"PASS","duration":16849655899,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.33.8.calico.base/NFS_deployment","result":"PASS","duration":17087361804,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.35.1.flannel.base","result":"PASS","duration":138589954127,"output":"        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.13.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.6-0\n        cluster.go:125: [init] Using Kubernetes version: v1.35.4\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:47:54.032338    1980 checks.go:906] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.158]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 501.136746ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.158:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.50861135s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 1.602850148s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 3.001875686s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 2av5k3.2joyzjxx61wa3dld\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.158:6443 --token 2av5k3.2joyzjxx61wa3dld \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:0b6c33ca64cf5c9dce5a4d620d260c82779f0c79f8ecf3505768c069e86845a6 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0421 10:48:21.665056    1621 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING ContainerRuntimeVersion]: You must update your container runtime to a version that supports the CRI method RuntimeConfig. Falling back to using cgroupDriver from kubelet config will be removed in 1.36. For more information, see https://git.k8s.io/enhancements/keps/sig-node/4033-group-driver-detection-over-cri\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.35.1.flannel.base/node_readiness (15.95s)\n    --- PASS: kubeadm.v1.35.1.flannel.base/nginx_deployment (10.99s)\n    --- PASS: kubeadm.v1.35.1.flannel.base/NFS_deployment (16.85s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.33.8.calico.base","result":"PASS","duration":156072756767,"output":"        cluster.go:125: I0421 10:47:22.662489    1815 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0421 10:47:30.856357    2034 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0421 10:47:30.963712    2034 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.157]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.00085555s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.157:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.168362777s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 1.289167047s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 3.002021631s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: qdbaph.vxawa15ozjx8vhvf\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.157:6443 --token qdbaph.vxawa15ozjx8vhvf \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:263d659e9ea79b14e99440abb7334b0fa70e8cba71f8d4280001ac745ca1837c \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: installation.operator.tigera.io/default created\n        cluster.go:125: apiserver.operator.tigera.io/default created\n        cluster.go:125: goldmane.operator.tigera.io/default created\n        cluster.go:125: whisker.operator.tigera.io/default created\n        cluster.go:125: W0421 10:48:14.193657    1645 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.calico.base/node_readiness (32.26s)\n    --- PASS: kubeadm.v1.33.8.calico.base/nginx_deployment (5.83s)\n    --- PASS: kubeadm.v1.33.8.calico.base/NFS_deployment (17.09s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"}],"result":"FAIL","platform":"qemu","version":"4081.3.7"}
