{"tests":[{"name":"cl.ignition.v1.btrfsroot","result":"PASS","duration":108748586209,"output":""},{"name":"coreos.tls.fetch-urls","result":"PASS","duration":255110669487,"output":""},{"name":"cl.tpm.root-cryptenroll-pcr-withupdate","result":"FAIL","duration":1388892398686,"output":"        cluster.go:125: New TPM2 token enrolled as key slot 1.\n        cluster.go:125: Wiped slot 2.\n        tpm.go:354: could not reboot machine: machine \"22689b37-6156-4995-9587-057b6ec002cf\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.4:22: connect: no route to host\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 22689b37-6156-4995-9587-057b6ec002cf console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 22689b37-6156-4995-9587-057b6ec002cf console\n"},{"name":"cl.sysext.fallbackdownload","result":"PASS","duration":176197140715,"output":"        update.go:664: Rebooting test machine\n"},{"name":"cl.cloudinit.basic","result":"PASS","duration":248224653118,"output":""},{"name":"kubeadm.v1.34.4.flannel.base","result":"FAIL","duration":1898235769457,"output":"        cluster.go:125: I0417 00:17:43.258929    2025 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0417 00:21:51.465328    2409 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.6]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 12.573135111s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.6:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 19.284874441s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 35.954834473s\n        cluster.go:125: E0417 00:30:17.019638    2409 request.go:1196] \"Unexpected error when reading response body\" err=\"context deadline exceeded (Client.Timeout or context cancellation while reading body)\"\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 12m35.313020862s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: xsxdiz.arazieta6ib347bk\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.6:6443 --token xsxdiz.arazieta6ib347bk \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:522f5d04f8d863098fa285ab668deeee5156df8c1bd37563f7133ba2e6509745 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0417 00:38:29.797047    1739 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: error: error execution phase preflight: couldn't validate the identity of the API Server: failed to request the cluster-info ConfigMap: Get \"https://10.0.0.6:6443/api/v1/namespaces/kube-public/configmaps/cluster-info?timeout=10s\": context deadline exceeded\n        cluster.go:125: To see the stack trace of this error execute with --v=5 or higher\n        kubeadm.go:197: unable to setup cluster: unable to run worker script: Process exited with status 1\n"},{"name":"kubeadm.v1.35.1.cilium.base","result":"FAIL","duration":1747317010080,"output":"        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.13.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.6-0\n        cluster.go:125: [init] Using Kubernetes version: v1.35.4\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.8]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 2.004329355s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.8:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 7.009493318s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 11.757730281s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 22.089083631s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: td72cb.m3t60ambr8kzzvse\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.8:6443 --token td72cb.m3t60ambr8kzzvse \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:69164211db2585c9d613cabcf0aed175862f1b00bb86b62c0f6071924b0d28df \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0417 00:41:39.174783    1827 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: error: error execution phase preflight: couldn't validate the identity of the API Server: failed to request the cluster-info ConfigMap: Get \"https://10.0.0.8:6443/api/v1/namespaces/kube-public/configmaps/cluster-info?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\n        cluster.go:125: To see the stack trace of this error execute with --v=5 or higher\n        kubeadm.go:197: unable to setup cluster: unable to run worker script: Process exited with status 1\n"},{"name":"packages","result":"FAIL","duration":287060882435,"output":"        harness.go:646: Cluster failed starting machines: machine \"a271470a-59be-4eac-9404-c49829389274\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine a271470a-59be-4eac-9404-c49829389274 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine a271470a-59be-4eac-9404-c49829389274 console\n"},{"name":"cl.verity/verify","result":"PASS","duration":13689692344,"output":"        cluster.go:125: Success\n"},{"name":"cl.toolbox.dnf-install","result":"PASS","duration":426782386656,"output":""},{"name":"kubeadm.v1.33.8.calico.base","result":"FAIL","duration":220368331918,"output":"        cluster.go:125: I0417 00:49:28.291979    2077 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0417 00:49:49.892148    2329 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.17]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.504473411s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.17:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 2.610668349s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 8.147341629s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 21.017740827s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 4hsvta.3hieji22cwn3330b\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.17:6443 --token 4hsvta.3hieji22cwn3330b \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:ff27f7456733af12c0b425ecf0172ca70a6feb1474ed10f955d963068801ef6c \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: error: .status.conditions accessor error: \u003cnil\u003e is of the type \u003cnil\u003e, expected []interface{}\n        kubeadm.go:197: unable to setup cluster: unable to run master script: Process exited with status 1\n"},{"name":"cl.ignition.v2.users","result":"PASS","duration":30100851633,"output":""},{"name":"cl.ignition.oem.regular.new","result":"FAIL","duration":138185231283,"output":"        oem.go:199: Couldn't reboot machine: machine \"b0ad4b32-89a7-48e4-9b32-db933ae899be\" failed basic checks: some systemd units failed:\n● ldconfig.service loaded failed failed Rebuild Dynamic Linker Cache\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39mldconfig.service\u001b[0m - Rebuild Dynamic Linker Cache.\r\r) on machine b0ad4b32-89a7-48e4-9b32-db933ae899be console\n"},{"name":"cl.verity/corruption","result":"PASS","duration":169856707630,"output":""},{"name":"cl.verity","result":"PASS","duration":221650256674,"output":"    --- PASS: cl.verity/verify (13.69s)\n            cluster.go:125: Success\n    --- PASS: cl.verity/corruption (169.86s)\n"},{"name":"coreos.ignition.resource.s3.versioned","result":"PASS","duration":43044245252,"output":""},{"name":"systemd.journal.remote","result":"PASS","duration":133571586793,"output":"        cluster.go:125: Running as unit: systemd-journal-remote-client.service; invocation ID: accb1b75754c4774af79bda21bd82174\n"},{"name":"cl.cloudinit.script","result":"PASS","duration":97141255865,"output":""},{"name":"cl.disk.raid0.root","result":"FAIL","duration":161955374685,"output":"        raid.go:231: could not reboot machine: machine \"1668e830-d9b3-4b10-a47c-1fbc11f440b8\" failed basic checks: some systemd units failed:\n● ldconfig.service loaded failed failed Rebuild Dynamic Linker Cache\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39mldconfig.service\u001b[0m - Rebuild Dynamic Linker Cache.\r\r) on machine 1668e830-d9b3-4b10-a47c-1fbc11f440b8 console\n"},{"name":"cl.ignition.v2_1.ext4checkexisting","result":"PASS","duration":95190117161,"output":""},{"name":"cl.update.reboot","result":"PASS","duration":104657131835,"output":""},{"name":"linux.ntp","result":"PASS","duration":34347344003,"output":""},{"name":"sysext.custom-docker.sysext","result":"PASS","duration":44622590593,"output":"        cluster.go:125: bash: line 1: docker: command not found\n        cluster.go:125: Cloning into 'sysext-bakery'...\n        cluster.go:125: Note: switching to '9850ffd5b2353f45a9b3bf4fb84f8138a149e3e7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at 9850ffd Merge pull request #31 from flatcar/t-lo/fix-docker-23-containerd-shim\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Download complete\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n"},{"name":"cl.ignition.oem.indirect","result":"PASS","duration":87492909805,"output":""},{"name":"cl.ignition.v1.sethostname","result":"PASS","duration":47388528835,"output":""},{"name":"coreos.ignition.groups","result":"PASS","duration":65717822467,"output":""},{"name":"sysext.simple","result":"PASS","duration":65555695499,"output":""},{"name":"docker.lib-coreos-dockerd-compat/docker-info","result":"PASS","duration":8690330713,"output":""},{"name":"docker.lib-coreos-dockerd-compat/resources","result":"PASS","duration":45407031774,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.2s done\n        cluster.go:125: #1 DONE 0.8s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B 0.1s done\n        cluster.go:125: #2 DONE 0.8s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 3.42MB 1.3s done\n        cluster.go:125: #3 DONE 1.7s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.7s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:105140fe9999585d821037397f718c1ead95e78b1c02d408c34d4fe97743ffe4 done\n        cluster.go:125: #5 naming to docker.io/library/sleep 0.0s done\n        cluster.go:125: #5 DONE 0.2s\n        cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n        cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n"},{"name":"coreos.locksmith.reboot","result":"PASS","duration":423052973020,"output":""},{"name":"cl.locksmith.cluster","result":"PASS","duration":524520896748,"output":""},{"name":"sysext.custom-oem","result":"PASS","duration":249995609526,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4459.2.4+nightly-20260416-2100\n        cluster.go:125: + version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-7171\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-7171\n        cluster.go:125: + sudo systemd-nspawn --console=pipe --bind-ro=/home/core/dev-container-script.sh --bind=/home/core/dev-container-workdir-7171:/work --image=flatcar_developer_container.bin --machine=flatcar-developer-container /bin/bash /home/core/dev-container-script.sh\n        cluster.go:125: + set -euo pipefail\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ VERSION_ID=4459.2.4\n        cluster.go:125: +++ BUILD_ID=nightly-20260416-2100\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4459.2.4+nightly-20260416-2100 (Oklo)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4459.2.4+nightly-20260416-2100:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4459.2.4+nightly-20260416-2100\n        cluster.go:125: + version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ VERSION_ID=4459.2.4\n        cluster.go:125: +++ BUILD_ID=nightly-20260416-2100\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4459.2.4+nightly-20260416-2100 (Oklo)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4459.2.4+nightly-20260416-2100:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4459.2.4\n        cluster.go:125: + version_id=4459.2.4\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64-usr\n        cluster.go:125: + board=amd64-usr\n        cluster.go:125: + mkdir -p /work/sysext_rootfs/usr/share/flatcar-sysext-kola-test\n        cluster.go:125: + echo 4459.2.4\n        cluster.go:125: + mkdir -p /work/sysext_rootfs/usr/lib/extension-release.d\n        cluster.go:125: + sysext_arch=x86-64\n        cluster.go:125: + [[ amd64-usr = \\a\\r\\m\\6\\4\\-\\u\\s\\r ]]\n        cluster.go:125: + metadata=('ID=flatcar' \"VERSION_ID=${version_id}\" \"ARCHITECTURE=${sysext_arch}\")\n        cluster.go:125: + metadata_file=/work/sysext_rootfs/usr/lib/extension-release.d/extension-release.oem-test\n        cluster.go:125: + printf '%s\\n' ID=flatcar VERSION_ID=4459.2.4 ARCHITECTURE=x86-64\n        cluster.go:125: + mksquashfs /work/sysext_rootfs /work/oem-test-4459.2.4+nightly-20260416-2100.raw -all-root\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ VERSION_ID=4459.2.4\n        cluster.go:125: +++ BUILD_ID=nightly-20260416-2100\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4459.2.4+nightly-20260416-2100 (Oklo)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4459.2.4+nightly-20260416-2100:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4459.2.4+nightly-20260416-2100\n        cluster.go:125: + version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: + sysext_file=/home/core/dev-container-workdir-7171/oem-test-4459.2.4+nightly-20260416-2100.raw\n        cluster.go:125: + [[ ! -e /home/core/dev-container-workdir-7171/oem-test-4459.2.4+nightly-20260416-2100.raw ]]\n        cluster.go:125: + [[ ! -e /oem/oem-release ]]\n        cluster.go:125: + printf '%s\\n' ID=test VERSION_ID=1.0.0 'NAME=testing stuff'\n        cluster.go:125: + sudo tee /oem/oem-release\n        cluster.go:125: + sudo mkdir -p /oem/sysext\n        cluster.go:125: + sudo mv /home/core/dev-container-workdir-7171/oem-test-4459.2.4+nightly-20260416-2100.raw /oem/sysext\n        cluster.go:125: + sudo touch /oem/sysext/active-oem-test\n        cluster.go:125: + sudo flatcar-reset --keep-machine-id --keep-paths /var/log\n        cluster.go:125: + set -euo pipefail\n        cluster.go:125: ++ systemd-sysext list --json=pretty\n        cluster.go:125: + list_out='[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"containerd-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/containerd-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776380118397796\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"docker-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/docker-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776380143171694\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"oem-test\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/oem-test.raw\",\n        cluster.go:125: \t\t\"time\" : 1776388014537247\n        cluster.go:125: \t}\n        cluster.go:125: ]'\n        cluster.go:125: ++ systemd-sysext status --json=pretty\n        cluster.go:125: + status_out='[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/opt\",\n        cluster.go:125: \t\t\"extensions\" : \"none\",\n        cluster.go:125: \t\t\"since\" : null\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/usr\",\n        cluster.go:125: \t\t\"extensions\" : [\n        cluster.go:125: \t\t\t\"containerd-flatcar\",\n        cluster.go:125: \t\t\t\"docker-flatcar\",\n        cluster.go:125: \t\t\t\"oem-test\"\n        cluster.go:125: \t\t],\n        cluster.go:125: \t\t\"since\" : 1776388054278000\n        cluster.go:125: \t}\n        cluster.go:125: ]'\n        cluster.go:125: + printf 'sysext list:\\n%s\\nsysext status:\\n%s\\n' '[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"containerd-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/containerd-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776380118397796\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"docker-flatcar\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/docker-flatcar.raw\",\n        cluster.go:125: \t\t\"time\" : 1776380143171694\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"name\" : \"oem-test\",\n        cluster.go:125: \t\t\"type\" : \"raw\",\n        cluster.go:125: \t\t\"path\" : \"/etc/extensions/oem-test.raw\",\n        cluster.go:125: \t\t\"time\" : 1776388014537247\n        cluster.go:125: \t}\n        cluster.go:125: ]' '[\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/opt\",\n        cluster.go:125: \t\t\"extensions\" : \"none\",\n        cluster.go:125: \t\t\"since\" : null\n        cluster.go:125: \t},\n        cluster.go:125: \t{\n        cluster.go:125: \t\t\"hierarchy\" : \"/usr\",\n        cluster.go:125: \t\t\"extensions\" : [\n        cluster.go:125: \t\t\t\"containerd-flatcar\",\n        cluster.go:125: \t\t\t\"docker-flatcar\",\n        cluster.go:125: \t\t\t\"oem-test\"\n        cluster.go:125: \t\t],\n        cluster.go:125: \t\t\"since\" : 1776388054278000\n        cluster.go:125: \t}\n        cluster.go:125: ]'\n        cluster.go:125: ++ jq '.[] | select(.name == \"oem-test\")'\n        cluster.go:125: + list_oem_test='{\n        cluster.go:125:   \"name\": \"oem-test\",\n        cluster.go:125:   \"type\": \"raw\",\n        cluster.go:125:   \"path\": \"/etc/extensions/oem-test.raw\",\n        cluster.go:125:   \"time\": 1776388014537247\n        cluster.go:125: }'\n        cluster.go:125: + [[ -z {\n        cluster.go:125:   \"name\": \"oem-test\",\n        cluster.go:125:   \"type\": \"raw\",\n        cluster.go:125:   \"path\": \"/etc/extensions/oem-test.raw\",\n        cluster.go:125:   \"time\": 1776388014537247\n        cluster.go:125: } ]]\n        cluster.go:125: ++ jq --raw-output .type\n        cluster.go:125: + oem_test_type=raw\n        cluster.go:125: + [[ raw != \\r\\a\\w ]]\n        cluster.go:125: ++ jq --raw-output .path\n        cluster.go:125: + oem_test_path=/etc/extensions/oem-test.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + oem_test_real_path=/oem/sysext/oem-test-4459.2.4+nightly-20260416-2100.raw\n        cluster.go:125: + [[ /etc/extensions/oem-test.raw != \\/\\o\\e\\m\\/\\s\\y\\s\\e\\x\\t\\/\\o\\e\\m\\-\\t\\e\\s\\t\\-\\4\\4\\5\\9\\.\\2\\.\\4\\+\\n\\i\\g\\h\\t\\l\\y\\-\\2\\0\\2\\6\\0\\4\\1\\6\\-\\2\\1\\0\\0\\.\\r\\a\\w ]]\n        cluster.go:125: + [[ /etc/extensions/oem-test.raw != \\/\\e\\t\\c\\/\\e\\x\\t\\e\\n\\s\\i\\o\\n\\s\\/\\o\\e\\m\\-\\t\\e\\s\\t\\.\\r\\a\\w ]]\n        cluster.go:125: ++ jq '.[] | select(.hierarchy == \"/usr\")'\n        cluster.go:125: + status_usr='{\n        cluster.go:125:   \"hierarchy\": \"/usr\",\n        cluster.go:125:   \"extensions\": [\n        cluster.go:125:     \"containerd-flatcar\",\n        cluster.go:125:     \"docker-flatcar\",\n        cluster.go:125:     \"oem-test\"\n        cluster.go:125:   ],\n        cluster.go:125:   \"since\": 1776388054278000\n        cluster.go:125: }'\n        cluster.go:125: + [[ -z {\n        cluster.go:125:   \"hierarchy\": \"/usr\",\n        cluster.go:125:   \"extensions\": [\n        cluster.go:125:     \"containerd-flatcar\",\n        cluster.go:125:     \"docker-flatcar\",\n        cluster.go:125:     \"oem-test\"\n        cluster.go:125:   ],\n        cluster.go:125:   \"since\": 1776388054278000\n        cluster.go:125: } ]]\n        cluster.go:125: ++ jq --raw-output '.extensions[] | select(. == \"oem-test\")'\n        cluster.go:125: + status_usr_extensions_oem_test=oem-test\n        cluster.go:125: + [[ oem-test != \\o\\e\\m\\-\\t\\e\\s\\t ]]\n        cluster.go:125: + f=/usr/share/flatcar-sysext-kola-test/file\n        cluster.go:125: + [[ ! -e /usr/share/flatcar-sysext-kola-test/file ]]\n        cluster.go:125: ++ cat /usr/share/flatcar-sysext-kola-test/file\n        cluster.go:125: + got=4459.2.4\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ VERSION_ID=4459.2.4\n        cluster.go:125: +++ BUILD_ID=nightly-20260416-2100\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4459.2.4+nightly-20260416-2100 (Oklo)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4459.2.4+nightly-20260416-2100:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4459.2.4\n        cluster.go:125: + ex=4459.2.4\n        cluster.go:125: + [[ 4459.2.4 != \\4\\4\\5\\9\\.\\2\\.\\4 ]]\n"},{"name":"linux.nfs.v3","result":"PASS","duration":130567194418,"output":"        nfs.go:80: NFS server booted.\n        nfs.go:85: Test file \"/tmp/tmp.Kx9zvKIDq3\" created on server.\n        nfs.go:125: NFS client booted.\n        nfs.go:133: Got NFS mount.\n"},{"name":"docker.btrfs-storage","result":"PASS","duration":32744218040,"output":""},{"name":"cl.network.wireguard","result":"PASS","duration":30757253722,"output":"        cluster.go:152: + ip --json address show kv0 | jq -r '.[] | .addr_info | .[] | select( .family == \"inet\") | .local'\n"},{"name":"cl.network.nftables","result":"PASS","duration":30685801853,"output":"        cluster.go:152: + sudo nft --json list ruleset | jq '.nftables[] | select(.rule) | .rule.expr[0].match.right'\n"},{"name":"cl.tpm.nonroot","result":"PASS","duration":100402127971,"output":""},{"name":"docker.lib-coreos-dockerd-compat/networks-reliably","result":"PASS","duration":645595584012,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.2s done\n        cluster.go:125: #1 DONE 0.6s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 5.26MB 0.1s done\n        cluster.go:125: #3 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 0.5s done\n        cluster.go:125: #5 writing image sha256:158767e627c574271b4d81e41c58e92aac3a54423f6817ade2bccc394e308f32 0.1s done\n        cluster.go:125: #5 naming to docker.io/library/ping\n        cluster.go:125: #5 naming to docker.io/library/ping 0.1s done\n        cluster.go:125: #5 DONE 1.0s\n"},{"name":"docker.lib-coreos-dockerd-compat/user-no-caps","result":"PASS","duration":3052438970,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n        cluster.go:125: #1 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B 0.0s done\n        cluster.go:125: #2 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 6.27MB 0.1s done\n        cluster.go:125: #3 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:f7b8249bc3dc4fb0f720ac49f805c9d7112e24543ea4f29234f225ad2c7938c7 done\n        cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n        cluster.go:125: #5 DONE 0.2s\n"},{"name":"docker.lib-coreos-dockerd-compat/ownership","result":"PASS","duration":14087625276,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/nginx\n        cluster.go:125: 9d11882d4b58: Pulling fs layer\n        cluster.go:125: 6a7080264fc9: Pulling fs layer\n        cluster.go:125: f176d5d8a1c1: Pulling fs layer\n        cluster.go:125: 63ef37274310: Pulling fs layer\n        cluster.go:125: 4744f2344932: Pulling fs layer\n        cluster.go:125: 39a24e95a1e9: Pulling fs layer\n        cluster.go:125: 400f8ac367e0: Pulling fs layer\n        cluster.go:125: 63ef37274310: Waiting\n        cluster.go:125: 4744f2344932: Waiting\n        cluster.go:125: 39a24e95a1e9: Waiting\n        cluster.go:125: 400f8ac367e0: Waiting\n        cluster.go:125: f176d5d8a1c1: Verifying Checksum\n        cluster.go:125: f176d5d8a1c1: Download complete\n        cluster.go:125: 63ef37274310: Verifying Checksum\n        cluster.go:125: 63ef37274310: Download complete\n        cluster.go:125: 9d11882d4b58: Verifying Checksum\n        cluster.go:125: 9d11882d4b58: Download complete\n        cluster.go:125: 6a7080264fc9: Verifying Checksum\n        cluster.go:125: 6a7080264fc9: Download complete\n        cluster.go:125: 4744f2344932: Download complete\n        cluster.go:125: 400f8ac367e0: Download complete\n        cluster.go:125: 39a24e95a1e9: Verifying Checksum\n        cluster.go:125: 39a24e95a1e9: Download complete\n        cluster.go:125: 9d11882d4b58: Pull complete\n        cluster.go:125: 6a7080264fc9: Pull complete\n        cluster.go:125: f176d5d8a1c1: Pull complete\n        cluster.go:125: 63ef37274310: Pull complete\n        cluster.go:125: 4744f2344932: Pull complete\n        cluster.go:125: 39a24e95a1e9: Pull complete\n        cluster.go:125: 400f8ac367e0: Pull complete\n        cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"docker.lib-coreos-dockerd-compat","result":"PASS","duration":772945906119,"output":"    --- PASS: docker.lib-coreos-dockerd-compat/docker-info (8.69s)\n    --- PASS: docker.lib-coreos-dockerd-compat/resources (45.41s)\n            cluster.go:125: #0 building with \"default\" instance using docker driver\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B 0.2s done\n            cluster.go:125: #1 DONE 0.8s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context:\n            cluster.go:125: #2 transferring context: 2B 0.1s done\n            cluster.go:125: #2 DONE 0.8s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 3.42MB 1.3s done\n            cluster.go:125: #3 DONE 1.7s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.7s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.1s done\n            cluster.go:125: #5 writing image sha256:105140fe9999585d821037397f718c1ead95e78b1c02d408c34d4fe97743ffe4 done\n            cluster.go:125: #5 naming to docker.io/library/sleep 0.0s done\n            cluster.go:125: #5 DONE 0.2s\n            cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n            cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n    --- PASS: docker.lib-coreos-dockerd-compat/networks-reliably (645.60s)\n            cluster.go:125: #0 building with \"default\" instance using docker driver\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B 0.2s done\n            cluster.go:125: #1 DONE 0.6s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context:\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.2s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 5.26MB 0.1s done\n            cluster.go:125: #3 DONE 0.2s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.2s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers\n            cluster.go:125: #5 exporting layers 0.5s done\n            cluster.go:125: #5 writing image sha256:158767e627c574271b4d81e41c58e92aac3a54423f6817ade2bccc394e308f32 0.1s done\n            cluster.go:125: #5 naming to docker.io/library/ping\n            cluster.go:125: #5 naming to docker.io/library/ping 0.1s done\n            cluster.go:125: #5 DONE 1.0s\n    --- PASS: docker.lib-coreos-dockerd-compat/user-no-caps (3.05s)\n            cluster.go:125: #0 building with \"default\" instance using docker driver\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n            cluster.go:125: #1 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B 0.0s done\n            cluster.go:125: #2 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 6.27MB 0.1s done\n            cluster.go:125: #3 DONE 0.2s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.1s done\n            cluster.go:125: #5 writing image sha256:f7b8249bc3dc4fb0f720ac49f805c9d7112e24543ea4f29234f225ad2c7938c7 done\n            cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n            cluster.go:125: #5 DONE 0.2s\n    --- PASS: docker.lib-coreos-dockerd-compat/ownership (14.09s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/nginx\n            cluster.go:125: 9d11882d4b58: Pulling fs layer\n            cluster.go:125: 6a7080264fc9: Pulling fs layer\n            cluster.go:125: f176d5d8a1c1: Pulling fs layer\n            cluster.go:125: 63ef37274310: Pulling fs layer\n            cluster.go:125: 4744f2344932: Pulling fs layer\n            cluster.go:125: 39a24e95a1e9: Pulling fs layer\n            cluster.go:125: 400f8ac367e0: Pulling fs layer\n            cluster.go:125: 63ef37274310: Waiting\n            cluster.go:125: 4744f2344932: Waiting\n            cluster.go:125: 39a24e95a1e9: Waiting\n            cluster.go:125: 400f8ac367e0: Waiting\n            cluster.go:125: f176d5d8a1c1: Verifying Checksum\n            cluster.go:125: f176d5d8a1c1: Download complete\n            cluster.go:125: 63ef37274310: Verifying Checksum\n            cluster.go:125: 63ef37274310: Download complete\n            cluster.go:125: 9d11882d4b58: Verifying Checksum\n            cluster.go:125: 9d11882d4b58: Download complete\n            cluster.go:125: 6a7080264fc9: Verifying Checksum\n            cluster.go:125: 6a7080264fc9: Download complete\n            cluster.go:125: 4744f2344932: Download complete\n            cluster.go:125: 400f8ac367e0: Download complete\n            cluster.go:125: 39a24e95a1e9: Verifying Checksum\n            cluster.go:125: 39a24e95a1e9: Download complete\n            cluster.go:125: 9d11882d4b58: Pull complete\n            cluster.go:125: 6a7080264fc9: Pull complete\n            cluster.go:125: f176d5d8a1c1: Pull complete\n            cluster.go:125: 63ef37274310: Pull complete\n            cluster.go:125: 4744f2344932: Pull complete\n            cluster.go:125: 39a24e95a1e9: Pull complete\n            cluster.go:125: 400f8ac367e0: Pull complete\n            cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"coreos.auth.verify","result":"PASS","duration":34661035928,"output":""},{"name":"cl.tpm.eventlog","result":"PASS","duration":31568695540,"output":"        cluster.go:125: WARN: Event 16 is unexectedly not extending either PCR 8, 9, 12 or 14\n"},{"name":"cl.basic/Script","result":"PASS","duration":297435064,"output":""},{"name":"cl.basic/PortSSH","result":"PASS","duration":104820830,"output":""},{"name":"cl.basic/Symlink","result":"PASS","duration":131147576,"output":""},{"name":"cl.basic/UpdateEngineKeys","result":"PASS","duration":159178945,"output":""},{"name":"cl.basic/ServicesActive","result":"PASS","duration":125623988,"output":""},{"name":"cl.basic/RandomUUID","result":"PASS","duration":127580903,"output":""},{"name":"cl.basic/Useradd","result":"PASS","duration":198947068,"output":""},{"name":"cl.basic/MachineID","result":"PASS","duration":146891733,"output":""},{"name":"cl.basic/CloudConfig","result":"PASS","duration":384135054,"output":""},{"name":"cl.basic/DbusPerms","result":"PASS","duration":172558270,"output":""},{"name":"cl.basic/SymlinkFlatcar","result":"PASS","duration":108808945,"output":""},{"name":"cl.basic/ReadOnly","result":"PASS","duration":101312354,"output":""},{"name":"cl.basic/Microcode","result":"PASS","duration":140946529,"output":""},{"name":"cl.basic","result":"PASS","duration":28896457773,"output":"    --- PASS: cl.basic/Script (0.30s)\n    --- PASS: cl.basic/PortSSH (0.10s)\n    --- PASS: cl.basic/Symlink (0.13s)\n    --- PASS: cl.basic/UpdateEngineKeys (0.16s)\n    --- PASS: cl.basic/ServicesActive (0.13s)\n    --- PASS: cl.basic/RandomUUID (0.13s)\n    --- PASS: cl.basic/Useradd (0.20s)\n    --- PASS: cl.basic/MachineID (0.15s)\n    --- PASS: cl.basic/CloudConfig (0.38s)\n    --- PASS: cl.basic/DbusPerms (0.17s)\n    --- PASS: cl.basic/SymlinkFlatcar (0.11s)\n    --- PASS: cl.basic/ReadOnly (0.10s)\n    --- PASS: cl.basic/Microcode (0.14s)\n"},{"name":"cl.tpm.root-cryptenroll-pcr-noupdate","result":"PASS","duration":211476574199,"output":""},{"name":"coreos.ignition.security.tls","result":"PASS","duration":82013415145,"output":""},{"name":"linux.nfs.v4","result":"PASS","duration":206979384259,"output":"        nfs.go:80: NFS server booted.\n        nfs.go:85: Test file \"/tmp/tmp.ijPJHbc6Yz\" created on server.\n        nfs.go:125: NFS client booted.\n        nfs.go:133: Got NFS mount.\n"},{"name":"kubeadm.v1.34.4.calico.base","result":"FAIL","duration":367023763001,"output":"        cluster.go:125: I0417 01:16:40.352592    2059 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0417 01:17:08.053155    2313 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.58]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 2.005394168s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.58:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 10.156693219s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 25.441909286s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 40.611702136s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 7u73wv.lkd58adgroqsda2n\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.58:6443 --token 7u73wv.lkd58adgroqsda2n \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:630f9590d3664538cd6d969809ee213c8f4a5219d949c151a20f4c5b0ff37eb7 \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: error: .status.conditions accessor error: \u003cnil\u003e is of the type \u003cnil\u003e, expected []interface{}\n        kubeadm.go:197: unable to setup cluster: unable to run master script: Process exited with status 1\n"},{"name":"kubeadm.v1.34.4.cilium.base/node_readiness","result":"PASS","duration":54661785086,"output":""},{"name":"kubeadm.v1.34.4.cilium.base/nginx_deployment","result":"PASS","duration":41169400427,"output":""},{"name":"kubeadm.v1.34.4.cilium.base/NFS_deployment","result":"PASS","duration":28411962493,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.34.4.cilium.base/IPSec_encryption","result":"PASS","duration":59509358443,"output":"        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"kubeadm.v1.34.4.cilium.base","result":"PASS","duration":582596355905,"output":"        cluster.go:125: I0417 01:14:37.332971    2050 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0417 01:15:47.952397    2354 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.55]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.506988261s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.55:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 6.770195869s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 8.39062343s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 13.007867921s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: t0yhzf.1lrsl30fonoji1st\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.55:6443 --token t0yhzf.1lrsl30fonoji1st \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:6061e364138358f7aa9d40433815981bf032773f606d004850c7d3a9fcab2989 \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0417 01:17:42.947055    1794 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.34.4.cilium.base/node_readiness (54.66s)\n    --- PASS: kubeadm.v1.34.4.cilium.base/nginx_deployment (41.17s)\n    --- PASS: kubeadm.v1.34.4.cilium.base/NFS_deployment (28.41s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n    --- PASS: kubeadm.v1.34.4.cilium.base/IPSec_encryption (59.51s)\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n"},{"name":"cl.ignition.symlink","result":"PASS","duration":34853908433,"output":"        cluster.go:152: + readlink /etc/localtime\n"},{"name":"cl.update.badverity","result":"PASS","duration":353817910526,"output":""},{"name":"systemd.sysusers.gshadow","result":"PASS","duration":63656162958,"output":""},{"name":"kubeadm.v1.35.1.flannel.base/node_readiness","result":"PASS","duration":60492050869,"output":""},{"name":"cl.etcd-member.v2-backup-restore","result":"PASS","duration":145768952460,"output":"        cluster.go:125: 2026-04-17T01:24:01Z\tinfo\tetcdutl/backup_command.go:216\tignoring EntryConfChange raft entry\n        cluster.go:125: 2026-04-17T01:24:01Z\tinfo\tetcdutl/backup_command.go:231\tignoring member attribute update on\t{\"entry\": \"Term:2 Index:3 Data:\\\"\\\\010\\\\202\\\\212\\\\246\\\\313\\\\220\\\\263\\\\347\\\\321\\\\340\\\\001\\\\022\\\\003PUT\\\\032\u0026/0/members/a89addc3880fe0a3/attributes\\\\\\\"R{\\\\\\\"name\\\\\\\":\\\\\\\"f2bd0d6e0a5b446a8041a26edd97cbcc\\\\\\\",\\\\\\\"clientURLs\\\\\\\":[\\\\\\\"http://10.0.0.65:2379\\\\\\\"]}(\\\\0002\\\\0008\\\\000H\\\\000P\\\\000X\\\\000`\\\\000h\\\\000p\\\\000x\\\\000\\\\200\\\\001\\\\000\\\" \", \"v2Req.Path\": \"/0/members/a89addc3880fe0a3/attributes\"}\n        cluster.go:125: 2026-04-17T01:24:01Z\tinfo\tetcdutl/backup_command.go:252\tignoring v3 raft entry\n        cluster.go:125: 2026-04-17T01:24:01Z\tinfo\tmembership/store.go:119\tTrimming membership information from the backend...\n"},{"name":"kubeadm.v1.35.1.flannel.base/nginx_deployment","result":"PASS","duration":21435189207,"output":""},{"name":"kubeadm.v1.35.1.flannel.base/NFS_deployment","result":"PASS","duration":36141352374,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.35.1.flannel.base","result":"PASS","duration":366845311101,"output":"        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.13.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.6-0\n        cluster.go:125: [init] Using Kubernetes version: v1.35.4\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.62]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 504.256839ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.62:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 8.536985746s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 11.209102167s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 14.002617248s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: s1vuu2.2vmv7nvidtkn2f4s\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.62:6443 --token s1vuu2.2vmv7nvidtkn2f4s \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:1e8b59a5325eb764040430795117dc7da6bf2f7582386c70fbcc2ec929a25dce \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0417 01:22:36.884689    1789 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.35.1.flannel.base/node_readiness (60.49s)\n    --- PASS: kubeadm.v1.35.1.flannel.base/nginx_deployment (21.44s)\n    --- PASS: kubeadm.v1.35.1.flannel.base/NFS_deployment (36.14s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"docker.network-openbsd-nc","result":"PASS","duration":62417747075,"output":"        docker.go:413: creating netcat containers\n        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 3.68MB 0.1s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:f7c527a76d95ce070c3b1e14f889b9dd26690ebe04c716d8329957a4aaf463a3 done\n        cluster.go:125: #5 naming to docker.io/library/netcat done\n        cluster.go:125: #5 DONE 0.1s\n        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B 0.0s done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 3.68MB 0.2s done\n        cluster.go:125: #3 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:3f35ad40537c374dfd6a30305b013d0f36426fa18151e72b7cca6ec27557c90a 0.0s done\n        cluster.go:125: #5 naming to docker.io/library/netcat 0.0s done\n        cluster.go:125: #5 DONE 0.1s\n"},{"name":"cl.ignition.oem.reuse","result":"PASS","duration":155880018470,"output":""},{"name":"cl.etcd-member.discovery","result":"PASS","duration":56219268389,"output":""},{"name":"cl.ignition.kargs","result":"PASS","duration":55967920520,"output":"        cluster.go:152: + cat /proc/cmdline\n"},{"name":"docker.userns","result":"PASS","duration":49005261449,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 3.47MB 0.0s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:433cc0747891dee314647fe874ea5b7826df6a09df4dea16168bcb5af11dc173 done\n        cluster.go:125: #5 naming to docker.io/library/userns-test done\n        cluster.go:125: #5 DONE 0.1s\n"},{"name":"cl.filesystem/deadlinks","result":"PASS","duration":2055797726,"output":""},{"name":"cl.filesystem/suid","result":"PASS","duration":240664591,"output":""},{"name":"cl.filesystem/sgid","result":"PASS","duration":184892973,"output":""},{"name":"cl.filesystem/writablefiles","result":"PASS","duration":186469110,"output":""},{"name":"cl.filesystem/writabledirs","result":"PASS","duration":170245937,"output":""},{"name":"cl.filesystem/stickydirs","result":"PASS","duration":166240965,"output":""},{"name":"cl.filesystem/blacklist","result":"PASS","duration":180812228,"output":""},{"name":"cl.filesystem","result":"PASS","duration":36937627298,"output":"    --- PASS: cl.filesystem/deadlinks (2.06s)\n    --- PASS: cl.filesystem/suid (0.24s)\n    --- PASS: cl.filesystem/sgid (0.18s)\n    --- PASS: cl.filesystem/writablefiles (0.19s)\n    --- PASS: cl.filesystem/writabledirs (0.17s)\n    --- PASS: cl.filesystem/stickydirs (0.17s)\n    --- PASS: cl.filesystem/blacklist (0.18s)\n"},{"name":"cl.ignition.v2.ext4root","result":"PASS","duration":35232377404,"output":""},{"name":"cl.etcd-member.etcdctlv3","result":"PASS","duration":40601441273,"output":"        cluster.go:125: {\"level\":\"info\",\"ts\":1776389211.4452465,\"caller\":\"snapshot/v3_snapshot.go:68\",\"msg\":\"created temporary db file\",\"path\":\"/tmp/tmp.u11oxf4kvV/snapshot.db.part\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776389211.4516327,\"logger\":\"client\",\"caller\":\"v3/maintenance.go:211\",\"msg\":\"opened snapshot stream; downloading\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776389211.4518552,\"caller\":\"snapshot/v3_snapshot.go:76\",\"msg\":\"fetching snapshot\",\"endpoint\":\"127.0.0.1:2379\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776389211.4560719,\"logger\":\"client\",\"caller\":\"v3/maintenance.go:219\",\"msg\":\"completed snapshot read; closing\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776389211.4561064,\"caller\":\"snapshot/v3_snapshot.go:91\",\"msg\":\"fetched snapshot\",\"endpoint\":\"127.0.0.1:2379\",\"size\":\"20 kB\",\"took\":\"now\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776389211.4561317,\"caller\":\"snapshot/v3_snapshot.go:100\",\"msg\":\"saved\",\"path\":\"/tmp/tmp.u11oxf4kvV/snapshot.db\"}\n        cluster.go:125: Deprecated: Use `etcdutl snapshot status` instead.\n"},{"name":"cl.swap_activation","result":"PASS","duration":34734513449,"output":""},{"name":"cl.ignition.v1.xfsroot","result":"PASS","duration":43234796432,"output":""},{"name":"misc.fips","result":"PASS","duration":51419064151,"output":"        cluster.go:125: Error setting digest\n        cluster.go:125: 808BFC1D457F0000:error:0308010C:digital envelope routines:inner_evp_generic_fetch:unsupported:../openssl-3.4.4/crypto/evp/evp_fetch.c:356:Global default library context, Algorithm (MD5 : 102), Properties ()\n        cluster.go:125: 808BFC1D457F0000:error:03000086:digital envelope routines:evp_md_init_internal:initialization error:../openssl-3.4.4/crypto/evp/digest.c:271:\n        cluster.go:152: + cat /proc/sys/crypto/fips_enabled\n"},{"name":"cl.install.cloudinit","result":"PASS","duration":35840425870,"output":""},{"name":"cl.disk.raid1.data","result":"PASS","duration":102430813417,"output":""},{"name":"docker.base/docker-info","result":"PASS","duration":15051941964,"output":""},{"name":"cl.ignition.v2_1.vfat","result":"PASS","duration":82799427501,"output":""},{"name":"docker.base/resources","result":"PASS","duration":105020588400,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.2s done\n        cluster.go:125: #1 DONE 0.7s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B 0.1s done\n        cluster.go:125: #2 DONE 0.4s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 3.42MB 2.6s done\n        cluster.go:125: #3 DONE 3.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.8s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 0.4s done\n        cluster.go:125: #5 writing image sha256:869543d0378ae80cee652c8200d0b68a1c39bd574b96b3d3cbb796f9b6322256 0.1s done\n        cluster.go:125: #5 naming to docker.io/library/sleep 0.1s done\n        cluster.go:125: #5 DONE 0.7s\n        cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n        cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n"},{"name":"kubeadm.v1.35.1.calico.base/node_readiness","result":"PASS","duration":152380869359,"output":""},{"name":"kubeadm.v1.35.1.calico.base/nginx_deployment","result":"PASS","duration":60365066050,"output":""},{"name":"kubeadm.v1.35.1.calico.base/NFS_deployment","result":"FAIL","duration":197357481626,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        kubeadm.go:264: nginx pod with NFS is not deployed: getting container status: Process exited with status 5\n"},{"name":"kubeadm.v1.35.1.calico.base","result":"FAIL","duration":679518611894,"output":"        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.13.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.6-0\n        cluster.go:125: [init] Using Kubernetes version: v1.35.4\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.82]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.001606513s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.82:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.507361607s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 2.385121863s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 4.502285573s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 17hkka.f4grn4dft65i8ubn\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.82:6443 --token 17hkka.f4grn4dft65i8ubn \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:fdb95f405cce722474bb3a42876b33c94d21677ee2027c26f5e2fe3940eaf106 \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: installation.operator.tigera.io/default created\n        cluster.go:125: apiserver.operator.tigera.io/default created\n        cluster.go:125: goldmane.operator.tigera.io/default created\n        cluster.go:125: whisker.operator.tigera.io/default created\n        cluster.go:125: W0417 01:30:32.341642    1816 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.35.1.calico.base/node_readiness (152.38s)\n    --- PASS: kubeadm.v1.35.1.calico.base/nginx_deployment (60.37s)\n    --- FAIL: kubeadm.v1.35.1.calico.base/NFS_deployment (197.36s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            kubeadm.go:264: nginx pod with NFS is not deployed: getting container status: Process exited with status 5\n"},{"name":"cl.sysext.boot","result":"PASS","duration":493826992031,"output":"        cluster.go:125: + sudo systemctl mask --now ensure-sysext\n        cluster.go:125: Created symlink '/etc/systemd/system/ensure-sysext.service' → '/dev/null'.\n        cluster.go:125: + sudo systemctl mask --now systemd-sysext\n        cluster.go:125: Created symlink '/etc/systemd/system/systemd-sysext.service' → '/dev/null'.\n        cluster.go:125: + sudo mkdir -p /etc/flatcar/sysext /etc/flatcar/oem-sysext /oem/sysext /etc/extensions\n        cluster.go:125: + sudo tee /oem/oem-release\n        cluster.go:125: + echo ID=test\n        cluster.go:125: + echo myext\n        cluster.go:125: + sudo tee /etc/flatcar/enabled-sysext.conf\n        cluster.go:125: + sudo touch /oem/sysext/active-oem-test /etc/flatcar/oem-sysext/oem-test-4459.2.4+nightly-20260416-2100.raw /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /etc/flatcar/sysext/flatcar-myext-4459.2.4+nightly-20260416-2100.raw /etc/flatcar/sysext/flatcar-myext-1.2.3.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /etc/extensions/oem-test.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/sysext/flatcar-myext-1.2.3.raw /etc/extensions/flatcar-myext.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4459.2.4+nightly-20260416-2100.raw = /oem/sysext/oem-test-4459.2.4+nightly-20260416-2100.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4459.2.4+nightly-20260416-2100.raw = /etc/flatcar/sysext/flatcar-myext-4459.2.4+nightly-20260416-2100.raw ']'\n        cluster.go:125: + sudo mv /oem/sysext/oem-test-4459.2.4+nightly-20260416-2100.raw /etc/flatcar/oem-sysext/\n        cluster.go:125: + sudo mv /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /oem/sysext/\n        cluster.go:125: + sudo ln -fs /oem/sysext/oem-test-1.2.3.raw /etc/extensions/oem-test.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/sysext/flatcar-myext-1.2.3.raw /etc/extensions/flatcar-myext.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4459.2.4+nightly-20260416-2100.raw = /oem/sysext/oem-test-4459.2.4+nightly-20260416-2100.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4459.2.4+nightly-20260416-2100.raw = /etc/flatcar/sysext/flatcar-myext-4459.2.4+nightly-20260416-2100.raw ']'\n        cluster.go:125: + '[' -e /etc/flatcar/oem-sysext/oem-test-1.2.3.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4459.2.4+nightly-20260416-2100.raw = /oem/sysext/oem-test-4459.2.4+nightly-20260416-2100.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4459.2.4+nightly-20260416-2100.raw = /etc/flatcar/sysext/flatcar-myext-4459.2.4+nightly-20260416-2100.raw ']'\n"},{"name":"cl.overlay.cleanup","result":"PASS","duration":79900262409,"output":""},{"name":"cl.ignition.v2_1.swap","result":"PASS","duration":61410216646,"output":""},{"name":"docker.containerd-restart","result":"PASS","duration":250333449497,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n"},{"name":"coreos.selinux.boolean","result":"FAIL","duration":289510091265,"output":"        harness.go:646: Cluster failed starting machines: machine \"bfebcf94-0d14-48de-a85f-31824b58ee91\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine bfebcf94-0d14-48de-a85f-31824b58ee91 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine bfebcf94-0d14-48de-a85f-31824b58ee91 console\n"},{"name":"cl.ignition.v1.ext4root","result":"FAIL","duration":495483971325,"output":"        harness.go:646: Cluster failed starting machines: machine \"a24bfd59-d810-4a0d-b8d2-081131f0b1c9\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine a24bfd59-d810-4a0d-b8d2-081131f0b1c9 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine a24bfd59-d810-4a0d-b8d2-081131f0b1c9 console\n"},{"name":"cl.tpm.root","result":"PASS","duration":463515944337,"output":""},{"name":"bpf.ig/ig","result":"PASS","duration":550880609653,"output":"        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n        cluster.go:125: time=\"2026-04-17T01:55:28Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Get \\\"http://%2Frun%2Fdocker.sock/v1.51/containers/json?all=1\\\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\"\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + docker info\n        cluster.go:125: + docker ps\n        cluster.go:125: + docker images\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n        cluster.go:125: + dig kinvolk.io\n        cluster.go:125: + dig flatcar.org\n        cluster.go:125: + dig stable.release.flatcar-linux.net\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"bpf.ig","result":"PASS","duration":553088914735,"output":"    --- PASS: bpf.ig/ig (550.88s)\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n            cluster.go:125: time=\"2026-04-17T01:55:28Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Get \\\"http://%2Frun%2Fdocker.sock/v1.51/containers/json?all=1\\\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\"\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + docker info\n            cluster.go:125: + docker ps\n            cluster.go:125: + docker images\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n            cluster.go:125: + dig kinvolk.io\n            cluster.go:125: + dig flatcar.org\n            cluster.go:125: + dig stable.release.flatcar-linux.net\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"cl.ignition.translation","result":"PASS","duration":106763204232,"output":"        cluster.go:152: + ip --json address show kola | jq -r '.[] | .addr_info | .[] | select( .family == \"inet\") | .local'\n        cluster.go:152: + cat /etc/systemd/network/00-dummy.network\n"},{"name":"coreos.ignition.once","result":"PASS","duration":85589436650,"output":""},{"name":"cl.ignition.v1.groups","result":"PASS","duration":35320796235,"output":""},{"name":"cl.ignition.luks","result":"PASS","duration":131889562485,"output":""},{"name":"cl.ignition.v1.once","result":"PASS","duration":79335729393,"output":""},{"name":"coreos.ignition.resource.remote","result":"PASS","duration":46223126618,"output":""},{"name":"cl.tang.nonroot","result":"PASS","duration":241640538869,"output":"        tang.go:148: Started tang on 10.0.0.1:37259\n"},{"name":"cl.users.shells","result":"PASS","duration":27675319127,"output":""},{"name":"kubeadm.v1.33.8.cilium.base/node_readiness","result":"PASS","duration":19482807585,"output":""},{"name":"coreos.ignition.systemd.enable-service","result":"PASS","duration":47705242633,"output":""},{"name":"kubeadm.v1.33.8.cilium.base/nginx_deployment","result":"PASS","duration":27850173977,"output":""},{"name":"kubeadm.v1.33.8.cilium.base/NFS_deployment","result":"PASS","duration":54592401400,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"systemd.journal.user","result":"PASS","duration":70540193008,"output":""},{"name":"docker.base/networks-reliably","result":"PASS","duration":2188802604855,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.2s done\n        cluster.go:125: #1 DONE 0.6s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B 0.0s done\n        cluster.go:125: #2 DONE 0.3s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 5.26MB 0.8s done\n        cluster.go:125: #3 DONE 1.3s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.4s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 0.3s done\n        cluster.go:125: #5 writing image sha256:94282535a891cfde2e373cf24d9feb51fa89f2030e0ae489c0a3a79a27ac602e 0.0s done\n        cluster.go:125: #5 naming to docker.io/library/ping 0.0s done\n        cluster.go:125: #5 DONE 0.6s\n"},{"name":"docker.base/user-no-caps","result":"PASS","duration":21191528775,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n        cluster.go:125: #1 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.2s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 6.27MB 1.5s done\n        cluster.go:125: #3 DONE 2.0s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.3s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 0.3s done\n        cluster.go:125: #5 writing image sha256:b8fd664374f82759df2f1077916e5f861fcc8e005c9a683a8784135163a7fc75 0.0s done\n        cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n        cluster.go:125: #5 DONE 0.4s\n"},{"name":"kubeadm.v1.33.8.cilium.base/IPSec_encryption","result":"FAIL","duration":85181888570,"output":"        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: Error: Unable to determine status:  timeout while waiting for status to become successful: context deadline exceeded\n        cluster.go:145: \"/opt/bin/cilium status --wait --wait-duration 1m\" failed: output \u001b[33m    /¯¯\\\n\u001b[36m /¯¯\u001b[33m\\__/\u001b[32m¯¯\\\u001b[0m    Cilium:         \u001b[31m2 errors\u001b[0m\n\u001b[36m \\__\u001b[31m/¯¯\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n\u001b[32m /¯¯\u001b[31m\\__/\u001b[35m¯¯\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n\u001b[32m \\__\u001b[34m/¯¯\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n\u001b[34m    \\__/\n\u001b[0m\nDaemonSet         cilium             Desired: 2, Ready: \u001b[33m1/2\u001b[0m, Available: \u001b[33m1/2\u001b[0m, Unavailable: \u001b[31m1/2\u001b[0m\nDeployment        cilium-operator    Desired: 1, Ready: \u001b[32m1/1\u001b[0m, Available: \u001b[32m1/1\u001b[0m\nContainers:       cilium             Running: \u001b[32m2\u001b[0m\n                  cilium-operator    Running: \u001b[32m1\u001b[0m\nCluster Pods:     5/5 managed by Cilium\nImage versions    cilium             quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: 2\n                  cilium-operator    quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: 1\nErrors:           cilium             cilium          1 pods of DaemonSet cilium are not ready\n                  cilium             cilium-z66wg    unable to retrieve cilium status: command terminated with exit code 1, status Process exited with status 1\n"},{"name":"kubeadm.v1.33.8.cilium.base","result":"FAIL","duration":565502371636,"output":"        cluster.go:125: I0417 02:03:13.734331    2101 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0417 02:05:12.312497    2453 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.104]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 3.525485977s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.104:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 4.691294544s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 7.694109129s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 14.508310604s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: mxoor9.5q543i67a3eo16cz\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.104:6443 --token mxoor9.5q543i67a3eo16cz \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:b7fc5e74496ee2dadefbbcf1241effb09aed6e13d8b40b45bf346c7845cb12a8 \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Containers:      cilium             \n        cluster.go:125:                  cilium-operator    \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        cluster.go:125: W0417 02:06:15.428441    1794 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.cilium.base/node_readiness (19.48s)\n    --- PASS: kubeadm.v1.33.8.cilium.base/nginx_deployment (27.85s)\n    --- PASS: kubeadm.v1.33.8.cilium.base/NFS_deployment (54.59s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n    --- FAIL: kubeadm.v1.33.8.cilium.base/IPSec_encryption (85.18s)\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n            cluster.go:125: Error: Unable to determine status:  timeout while waiting for status to become successful: context deadline exceeded\n            cluster.go:145: \"/opt/bin/cilium status --wait --wait-duration 1m\" failed: output \u001b[33m    /¯¯\\\n    \u001b[36m /¯¯\u001b[33m\\__/\u001b[32m¯¯\\\u001b[0m    Cilium:         \u001b[31m2 errors\u001b[0m\n    \u001b[36m \\__\u001b[31m/¯¯\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n    \u001b[32m /¯¯\u001b[31m\\__/\u001b[35m¯¯\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n    \u001b[32m \\__\u001b[34m/¯¯\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n    \u001b[34m    \\__/\n    \u001b[0m\n    DaemonSet         cilium             Desired: 2, Ready: \u001b[33m1/2\u001b[0m, Available: \u001b[33m1/2\u001b[0m, Unavailable: \u001b[31m1/2\u001b[0m\n    Deployment        cilium-operator    Desired: 1, Ready: \u001b[32m1/1\u001b[0m, Available: \u001b[32m1/1\u001b[0m\n    Containers:       cilium             Running: \u001b[32m2\u001b[0m\n                      cilium-operator    Running: \u001b[32m1\u001b[0m\n    Cluster Pods:     5/5 managed by Cilium\n    Image versions    cilium             quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: 2\n                      cilium-operator    quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: 1\n    Errors:           cilium             cilium          1 pods of DaemonSet cilium are not ready\n                      cilium             cilium-z66wg    unable to retrieve cilium status: command terminated with exit code 1, status Process exited with status 1\n"},{"name":"cl.network.iptables","result":"PASS","duration":83902855595,"output":"        cluster.go:152: + sudo nft --json list ruleset | jq '.nftables[] | select(.rule) | .rule.expr[0].match.right'\n"},{"name":"cl.ignition.instantiated.enable-unit","result":"PASS","duration":33142618936,"output":""},{"name":"docker.base/ownership","result":"PASS","duration":225295895200,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/nginx\n        cluster.go:125: 9d11882d4b58: Pulling fs layer\n        cluster.go:125: 6a7080264fc9: Pulling fs layer\n        cluster.go:125: f176d5d8a1c1: Pulling fs layer\n        cluster.go:125: 63ef37274310: Pulling fs layer\n        cluster.go:125: 4744f2344932: Pulling fs layer\n        cluster.go:125: 39a24e95a1e9: Pulling fs layer\n        cluster.go:125: 400f8ac367e0: Pulling fs layer\n        cluster.go:125: 4744f2344932: Waiting\n        cluster.go:125: 39a24e95a1e9: Waiting\n        cluster.go:125: 400f8ac367e0: Waiting\n        cluster.go:125: 63ef37274310: Waiting\n        cluster.go:125: f176d5d8a1c1: Verifying Checksum\n        cluster.go:125: f176d5d8a1c1: Download complete\n        cluster.go:125: 63ef37274310: Verifying Checksum\n        cluster.go:125: 63ef37274310: Download complete\n        cluster.go:125: 4744f2344932: Verifying Checksum\n        cluster.go:125: 4744f2344932: Download complete\n        cluster.go:125: 39a24e95a1e9: Verifying Checksum\n        cluster.go:125: 39a24e95a1e9: Download complete\n        cluster.go:125: 400f8ac367e0: Verifying Checksum\n        cluster.go:125: 400f8ac367e0: Download complete\n        cluster.go:125: 6a7080264fc9: Verifying Checksum\n        cluster.go:125: 6a7080264fc9: Download complete\n        cluster.go:125: 9d11882d4b58: Pull complete\n        cluster.go:125: 6a7080264fc9: Pull complete\n        cluster.go:125: f176d5d8a1c1: Pull complete\n        cluster.go:125: 63ef37274310: Pull complete\n        cluster.go:125: 4744f2344932: Pull complete\n        cluster.go:125: 39a24e95a1e9: Pull complete\n        cluster.go:125: 400f8ac367e0: Pull complete\n        cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"cl.flannel.vxlan","result":"PASS","duration":194711305442,"output":"        flannel.go:121: ping from ecbe5a81-d6a5-48a3-b876-91404a8315a1(10.254.78.0) to c867ba37-56e7-4dc3-895f-5f6665b4ac49(10.254.94.0)\n"},{"name":"docker.base","result":"PASS","duration":2681288315156,"output":"    --- PASS: docker.base/docker-info (15.05s)\n    --- PASS: docker.base/resources (105.02s)\n            cluster.go:125: #0 building with \"default\" instance using docker driver\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B 0.2s done\n            cluster.go:125: #1 DONE 0.7s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context:\n            cluster.go:125: #2 transferring context: 2B 0.1s done\n            cluster.go:125: #2 DONE 0.4s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 3.42MB 2.6s done\n            cluster.go:125: #3 DONE 3.1s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.8s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers\n            cluster.go:125: #5 exporting layers 0.4s done\n            cluster.go:125: #5 writing image sha256:869543d0378ae80cee652c8200d0b68a1c39bd574b96b3d3cbb796f9b6322256 0.1s done\n            cluster.go:125: #5 naming to docker.io/library/sleep 0.1s done\n            cluster.go:125: #5 DONE 0.7s\n            cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n            cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n    --- PASS: docker.base/networks-reliably (2188.80s)\n            cluster.go:125: #0 building with \"default\" instance using docker driver\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B 0.2s done\n            cluster.go:125: #1 DONE 0.6s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context:\n            cluster.go:125: #2 transferring context: 2B 0.0s done\n            cluster.go:125: #2 DONE 0.3s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 5.26MB 0.8s done\n            cluster.go:125: #3 DONE 1.3s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.4s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers\n            cluster.go:125: #5 exporting layers 0.3s done\n            cluster.go:125: #5 writing image sha256:94282535a891cfde2e373cf24d9feb51fa89f2030e0ae489c0a3a79a27ac602e 0.0s done\n            cluster.go:125: #5 naming to docker.io/library/ping 0.0s done\n            cluster.go:125: #5 DONE 0.6s\n    --- PASS: docker.base/user-no-caps (21.19s)\n            cluster.go:125: #0 building with \"default\" instance using docker driver\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n            cluster.go:125: #1 DONE 0.2s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context:\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.2s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 6.27MB 1.5s done\n            cluster.go:125: #3 DONE 2.0s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.3s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers\n            cluster.go:125: #5 exporting layers 0.3s done\n            cluster.go:125: #5 writing image sha256:b8fd664374f82759df2f1077916e5f861fcc8e005c9a683a8784135163a7fc75 0.0s done\n            cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n            cluster.go:125: #5 DONE 0.4s\n    --- PASS: docker.base/ownership (225.30s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/nginx\n            cluster.go:125: 9d11882d4b58: Pulling fs layer\n            cluster.go:125: 6a7080264fc9: Pulling fs layer\n            cluster.go:125: f176d5d8a1c1: Pulling fs layer\n            cluster.go:125: 63ef37274310: Pulling fs layer\n            cluster.go:125: 4744f2344932: Pulling fs layer\n            cluster.go:125: 39a24e95a1e9: Pulling fs layer\n            cluster.go:125: 400f8ac367e0: Pulling fs layer\n            cluster.go:125: 4744f2344932: Waiting\n            cluster.go:125: 39a24e95a1e9: Waiting\n            cluster.go:125: 400f8ac367e0: Waiting\n            cluster.go:125: 63ef37274310: Waiting\n            cluster.go:125: f176d5d8a1c1: Verifying Checksum\n            cluster.go:125: f176d5d8a1c1: Download complete\n            cluster.go:125: 63ef37274310: Verifying Checksum\n            cluster.go:125: 63ef37274310: Download complete\n            cluster.go:125: 4744f2344932: Verifying Checksum\n            cluster.go:125: 4744f2344932: Download complete\n            cluster.go:125: 39a24e95a1e9: Verifying Checksum\n            cluster.go:125: 39a24e95a1e9: Download complete\n            cluster.go:125: 400f8ac367e0: Verifying Checksum\n            cluster.go:125: 400f8ac367e0: Download complete\n            cluster.go:125: 6a7080264fc9: Verifying Checksum\n            cluster.go:125: 6a7080264fc9: Download complete\n            cluster.go:125: 9d11882d4b58: Pull complete\n            cluster.go:125: 6a7080264fc9: Pull complete\n            cluster.go:125: f176d5d8a1c1: Pull complete\n            cluster.go:125: 63ef37274310: Pull complete\n            cluster.go:125: 4744f2344932: Pull complete\n            cluster.go:125: 39a24e95a1e9: Pull complete\n            cluster.go:125: 400f8ac367e0: Pull complete\n            cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"cl.ignition.v2.xfsroot","result":"PASS","duration":182037946430,"output":""},{"name":"cl.ignition.partition_on_boot_disk","result":"PASS","duration":27164678632,"output":"        filesystem.go:487: lsblk -f:\nNAME    FSTYPE   FSVER LABEL      UUID                                 FSAVAIL FSUSE% MOUNTPOINTS\nloop2   squashfs 4.0                                                                  \nloop3   squashfs 4.0                                                                  \nsr0                                                                                   \nvda                                                                                   \n|-vda1  vfat     FAT32 EFI-SYSTEM 6E81-8997                                           \n|-vda2                                                                                \n|-vda3  btrfs                     d1542dca-1171-4bcf-9aae-d85dd05fe503                \n| `-usr btrfs                     d1542dca-1171-4bcf-9aae-d85dd05fe503    1.2G     0% /usr\n|-vda4                                                                                \n|-vda6  btrfs          OEM        aa52e89c-0ed3-4175-9a87-dc7b421a671a    123M     1% /oem\n|-vda7                                                                                \n|-vda9  ext4     1.0   ROOT       ee420a69-62b9-42f4-84c7-ea3f2d87c569    5.9G     0% /\n`-vda10 xfs            VAR        e69420c3-eb4f-48c8-befd-3fcf6d92063c    5.2G     3% /var\n        filesystem.go:489: findmnt:\nTARGET                                                        SOURCE          FSTYPE     OPTIONS\n/                                                             /dev/vda9       ext4       rw,relatime,seclabel\n|-/boot                                                       systemd-1       autofs     rw,relatime,fd=33,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=5958\n|-/media                                                      tmpfs           tmpfs      rw,nosuid,nodev,noexec,relatime,seclabel\n|-/tmp                                                        tmpfs           tmpfs      rw,nosuid,nodev,seclabel,size=1236204k,nr_inodes=1048576\n|-/dev                                                        devtmpfs        devtmpfs   rw,nosuid,seclabel,size=4096k,nr_inodes=303205,mode=755\n| |-/dev/hugepages                                            hugetlbfs       hugetlbfs  rw,nosuid,nodev,relatime,seclabel,pagesize=2M\n| |-/dev/mqueue                                               mqueue          mqueue     rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/dev/shm                                                  tmpfs           tmpfs      rw,nosuid,nodev,seclabel\n| `-/dev/pts                                                  devpts          devpts     rw,nosuid,noexec,relatime,seclabel,gid=5,mode=620,ptmxmode=000\n|-/sys                                                        sysfs           sysfs      rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/fs/selinux                                           selinuxfs       selinuxfs  rw,nosuid,noexec,relatime\n| |-/sys/kernel/debug                                         debugfs         debugfs    rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/kernel/tracing                                       tracefs         tracefs    rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/kernel/security                                      securityfs      securityfs rw,nosuid,nodev,noexec,relatime\n| |-/sys/fs/cgroup                                            cgroup2         cgroup2    rw,nosuid,nodev,noexec,relatime,seclabel,nsdelegate,memory_recursiveprot\n| |-/sys/fs/pstore                                            pstore          pstore     rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/fs/bpf                                               bpf             bpf        rw,nosuid,nodev,noexec,relatime,mode=700\n| |-/sys/fs/fuse/connections                                  fusectl         fusectl    rw,nosuid,nodev,noexec,relatime\n| `-/sys/kernel/config                                        configfs        configfs   rw,nosuid,nodev,noexec,relatime\n|-/proc                                                       proc            proc       rw,nosuid,nodev,noexec,relatime\n| `-/proc/sys/fs/binfmt_misc                                  systemd-1       autofs     rw,relatime,fd=34,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=5962\n|-/run                                                        tmpfs           tmpfs      rw,nosuid,nodev,seclabel,size=494484k,nr_inodes=819200,mode=755\n| |-/run/user/500                                             tmpfs           tmpfs      rw,nosuid,nodev,relatime,seclabel,size=247240k,nr_inodes=61810,mode=700,uid=500,gid=500\n| |-/run/credentials/systemd-journald.service                 tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-network-generator.service        tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-udev-load-credentials.service    tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-sysctl.service                   tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-tmpfiles-setup-dev-early.service tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-sysusers.service                 tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-tmpfiles-setup-dev.service       tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-networkd.service                 tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-vconsole-setup.service           tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-tmpfiles-setup.service           tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/systemd-resolved.service                 tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| |-/run/credentials/serial-getty@ttyS0.service               tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n| `-/run/credentials/getty@tty1.service                       tmpfs           tmpfs      ro,nosuid,nodev,noexec,relatime,nosymfollow,seclabel,size=1024k,nr_inodes=1024,mode=700,noswap\n|-/var                                                        /dev/vda10      xfs        rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota\n|-/usr                                                        /dev/mapper/usr btrfs      ro,relatime,seclabel,rescue=nologreplay,space_cache=v2,subvolid=5,subvol=/\n| `-/usr                                                      sysext          overlay    ro,nodev,relatime,seclabel,lowerdir=/run/systemd/sysext/meta/usr:/run/systemd/sysext/extensions/docker-flatcar/usr:/run/systemd/sysext/extensions/containerd-flatcar/usr:/usr\n|-/etc                                                        overlay         overlay    rw,noatime,seclabel,lowerdir=/sysroot/usr/share/flatcar/etc,upperdir=/sysroot/etc,workdir=/sysroot/.etc-work,uuid=on,metacopy=off\n`-/oem                                                        /dev/vda6       btrfs      rw,nodev,relatime,seclabel,discard=async,space_cache=v2,subvolid=5,subvol=/\n        cluster.go:152: + findmnt /var\n        cluster.go:152: + cat /var/hello\n"},{"name":"coreos.ignition.sethostname","result":"PASS","duration":45633804093,"output":""},{"name":"sysext.disable-docker","result":"PASS","duration":34033932833,"output":""},{"name":"cl.internet/NTPDate","result":"PASS","duration":13881666762,"output":""},{"name":"cl.internet/UpdateEngine","result":"PASS","duration":5460227871,"output":""},{"name":"cl.internet/DockerPing","result":"PASS","duration":25186935789,"output":""},{"name":"cl.internet/DockerEcho","result":"PASS","duration":4076669762,"output":""},{"name":"cl.internet","result":"PASS","duration":145578389336,"output":"    --- PASS: cl.internet/NTPDate (13.88s)\n    --- PASS: cl.internet/UpdateEngine (5.46s)\n    --- PASS: cl.internet/DockerPing (25.19s)\n    --- PASS: cl.internet/DockerEcho (4.08s)\n"},{"name":"coreos.update.badusr","result":"PASS","duration":402815801029,"output":""},{"name":"cl.ignition.oem.indirect.new","result":"PASS","duration":60475636851,"output":""},{"name":"coreos.locksmith.tls","result":"PASS","duration":109318739510,"output":""},{"name":"cl.ignition.v1.users","result":"PASS","duration":93992609548,"output":""},{"name":"docker.enable-service.sysext","result":"PASS","duration":133560872650,"output":"        cluster.go:152: + systemctl is-enabled docker\n"},{"name":"cl.disk.raid1.root","result":"FAIL","duration":81548577755,"output":"        raid.go:231: could not reboot machine: machine \"a710e272-c525-454c-80b8-95848308aa19\" failed basic checks: some systemd units failed:\n● ldconfig.service loaded failed failed Rebuild Dynamic Linker Cache\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39mldconfig.service\u001b[0m - Rebuild Dynamic Linker Cache.\r\r) on machine a710e272-c525-454c-80b8-95848308aa19 console\n"},{"name":"sysext.disable-containerd","result":"PASS","duration":27409865754,"output":""},{"name":"cl.ignition.v2.btrfsroot","result":"PASS","duration":35108919678,"output":""},{"name":"cl.network.initramfs.second-boot","result":"PASS","duration":65875652400,"output":""},{"name":"coreos.selinux.enforce","result":"PASS","duration":131103562194,"output":""},{"name":"cl.cloudinit.multipart-mime","result":"PASS","duration":47708994637,"output":""},{"name":"cl.tang.root","result":"PASS","duration":112598910149,"output":"        tang.go:148: Started tang on 10.0.0.1:37963\n"},{"name":"devcontainer.docker","result":"PASS","duration":1338387261650,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4459.2.4+nightly-20260416-2100\n        cluster.go:125: + version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + ARCH=amd64\n        cluster.go:125: + VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs amd64 4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: + EXPECTED_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-7885\n        cluster.go:125: + USR_SRC_DIR=/home/core/dev-container-workdir-7885/src\n        cluster.go:125: + VAR_TMP_DIR=/home/core/dev-container-workdir-7885/tmp\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-7885/src /home/core/dev-container-workdir-7885/tmp\n        cluster.go:125: ++ parted flatcar_developer_container.bin unit b print\n        cluster.go:125: ++ grep Start --after-context=1\n        cluster.go:125: ++ tail --lines=1\n        cluster.go:125: ++ head --bytes=-2\n        cluster.go:125: ++ awk '{ print $2 }'\n        cluster.go:125: + offset=2097152\n        cluster.go:125: + mkdir root\n        cluster.go:125: + sudo mount -o loop,ro,offset=2097152 flatcar_developer_container.bin root\n        cluster.go:125: + sudo tar -C root -czf dev-container-image.tar.gz .\n        cluster.go:125: + sudo umount root\n        cluster.go:125: + rm -f flatcar_developer_container.bin\n        cluster.go:125: + docker import dev-container-image.tar.gz dev-container:42\n        cluster.go:125: + sudo rm -f dev-container-image.tar.gz\n        cluster.go:125: + restorecon /home/core/dev-container-script\n        cluster.go:125: + docker run --log-driver=journald --env PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs --env EXPECTED_VERSION=4459.2.4+nightly-20260416-2100 --mount type=bind,source=/lib/modules,target=/lib/modules,readonly=true --mount type=bind,source=/home/core/dev-container-script,target=/home/core/dev-container-script,readonly=true --mount type=bind,source=/home/core/dev-container-workdir-7885/src,target=/usr/src --mount type=bind,source=/home/core/dev-container-workdir-7885/tmp,target=/var/tmp dev-container:42 /bin/bash /home/core/dev-container-script\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + [[ 4459.2.4+nightly-20260416-2100 != \\4\\4\\5\\9\\.\\2\\.\\4\\+\\n\\i\\g\\h\\t\\l\\y\\-\\2\\0\\2\\6\\0\\4\\1\\6\\-\\2\\1\\0\\0 ]]\n        cluster.go:125: + export PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: + export 'FEATURES=-ipc-sandbox -network-sandbox'\n        cluster.go:125: + FEATURES='-ipc-sandbox -network-sandbox'\n        cluster.go:125: + emerge-gitclone\n        cluster.go:125: Cloning into '/var/lib/portage/scripts'...\n        cluster.go:125: Updating files:  37% (5094/13423)\rUpdating files:  38% (5101/13423)\rUpdating files:  39% (5235/13423)\rUpdating files:  40% (5370/13423)\rUpdating files:  41% (5504/13423)\rUpdating files:  42% (5638/13423)\rUpdating files:  43% (5772/13423)\rUpdating files:  44% (5907/13423)\rUpdating files:  45% (6041/13423)\rUpdating files:  46% (6175/13423)\rUpdating files:  47% (6309/13423)\rUpdating files:  48% (6444/13423)\rUpdating files:  49% (6578/13423)\rUpdating files:  50% (6712/13423)\rUpdating files:  51% (6846/13423)\rUpdating files:  52% (6980/13423)\rUpdating files:  53% (7115/13423)\rUpdating files:  53% (7122/13423)\rUpdating files:  54% (7249/13423)\rUpdating files:  55% (7383/13423)\rUpdating files:  56% (7517/13423)\rUpdating files:  57% (7652/13423)\rUpdating files:  58% (7786/13423)\rUpdating files:  59% (7920/13423)\rUpdating files:  60% (8054/13423)\rUpdating files:  61% (8189/13423)\rUpdating files:  62% (8323/13423)\rUpdating files:  63% (8457/13423)\rUpdating files:  64% (8591/13423)\rUpdating files:  65% (8725/13423)\rUpdating files:  66% (8860/13423)\rUpdating files:  67% (8994/13423)\rUpdating files:  68% (9128/13423)\rUpdating files:  69% (9262/13423)\rUpdating files:  70% (9397/13423)\rUpdating files:  71% (9531/13423)\rUpdating files:  72% (9665/13423)\rUpdating files:  73% (9799/13423)\rUpdating files:  74% (9934/13423)\rUpdating files:  75% (10068/13423)\rUpdating files:  76% (10202/13423)\rUpdating files:  77% (10336/13423)\rUpdating files:  78% (10470/13423)\rUpdating files:  79% (10605/13423)\rUpdating files:  80% (10739/13423)\rUpdating files:  81% (10873/13423)\rUpdating files:  82% (11007/13423)\rUpdating files:  83% (11142/13423)\rUpdating files:  84% (11276/13423)\rUpdating files:  85% (11410/13423)\rUpdating files:  86% (11544/13423)\rUpdating files:  87% (11679/13423)\rUpdating files:  88% (11813/13423)\rUpdating files:  89% (11947/13423)\rUpdating files:  89% (12019/13423)\rUpdating files:  90% (12081/13423)\rUpdating files:  91% (12215/13423)\rUpdating files:  92% (12350/13423)\rUpdating files:  93% (12484/13423)\rUpdating files:  94% (12618/13423)\rUpdating files:  95% (12752/13423)\rUpdating files:  96% (12887/13423)\rUpdating files:  97% (13021/13423)\rUpdating files:  98% (13155/13423)\rUpdating files:  99% (13289/13423)\rUpdating files: 100% (13423/13423)\rUpdating files: 100% (13423/13423), done.\n        cluster.go:125: Note: switching to 'stable-4459.2.4-nightly-20260416-2100'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at 14b6d8f376 New version: stable-4459.2.4-nightly-20260416-2100\n        cluster.go:125: + emerge --getbinpkg --verbose coreos-sources\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: + zcat /proc/config.gz\n        cluster.go:125: ++ nproc\n        cluster.go:125: + exec make -C /usr/src/linux -j4 modules_prepare V=1\n        cluster.go:125: \n        cluster.go:125: Error in reading or end of file.\n"},{"name":"cl.tpm.root-cryptenroll","result":"PASS","duration":79370394630,"output":""},{"name":"cl.network.listeners","result":"PASS","duration":26460507057,"output":""},{"name":"docker.selinux","result":"PASS","duration":35194395314,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Download complete\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: sh: can't create /opt/hello: Permission denied\n"},{"name":"cl.disk.raid0.data","result":"PASS","duration":64412897106,"output":""},{"name":"cl.ignition.oem.regular","result":"PASS","duration":54057828799,"output":""},{"name":"cl.ignition.oem.wipe","result":"PASS","duration":32171266654,"output":""},{"name":"cl.osreset.ignition-rerun","result":"PASS","duration":58454713281,"output":""},{"name":"cl.omaha.ping","result":"PASS","duration":25658858888,"output":""},{"name":"devcontainer.systemd-nspawn","result":"PASS","duration":1299638126254,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4459.2.4+nightly-20260416-2100\n        cluster.go:125: + version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4459.2.4+nightly-20260416-2100/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + ARCH=amd64\n        cluster.go:125: + VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs amd64 4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: + EXPECTED_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-19950\n        cluster.go:125: + USR_SRC_DIR=/home/core/dev-container-workdir-19950/src\n        cluster.go:125: + VAR_TMP_DIR=/home/core/dev-container-workdir-19950/tmp\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-19950/src /home/core/dev-container-workdir-19950/tmp\n        cluster.go:125: + sudo systemd-nspawn --console=pipe --setenv=PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs --setenv=EXPECTED_VERSION=4459.2.4+nightly-20260416-2100 --bind-ro=/lib/modules --bind-ro=/home/core/dev-container-script --bind=/home/core/dev-container-workdir-19950/src:/usr/src --bind=/home/core/dev-container-workdir-19950/tmp:/var/tmp --image=flatcar_developer_container.bin --machine=flatcar-developer-container /bin/bash /home/core/dev-container-script\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4459.2.4+nightly-20260416-2100\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + [[ 4459.2.4+nightly-20260416-2100 != \\4\\4\\5\\9\\.\\2\\.\\4\\+\\n\\i\\g\\h\\t\\l\\y\\-\\2\\0\\2\\6\\0\\4\\1\\6\\-\\2\\1\\0\\0 ]]\n        cluster.go:125: + export PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4459.2.4+nightly-20260416-2100/pkgs\n        cluster.go:125: + export 'FEATURES=-ipc-sandbox -network-sandbox'\n        cluster.go:125: + FEATURES='-ipc-sandbox -network-sandbox'\n        cluster.go:125: + emerge-gitclone\n        cluster.go:125: Cloning into '/var/lib/portage/scripts'...\n        cluster.go:125: Updating files:  13% (1778/13423)\rUpdating files:  14% (1880/13423)\rUpdating files:  15% (2014/13423)\rUpdating files:  16% (2148/13423)\rUpdating files:  17% (2282/13423)\rUpdating files:  17% (2342/13423)\rUpdating files:  18% (2417/13423)\rUpdating files:  19% (2551/13423)\rUpdating files:  19% (2643/13423)\rUpdating files:  20% (2685/13423)\rUpdating files:  21% (2819/13423)\rUpdating files:  22% (2954/13423)\rUpdating files:  23% (3088/13423)\rUpdating files:  23% (3161/13423)\rUpdating files:  24% (3222/13423)\rUpdating files:  25% (3356/13423)\rUpdating files:  26% (3490/13423)\rUpdating files:  27% (3625/13423)\rUpdating files:  28% (3759/13423)\rUpdating files:  28% (3810/13423)\rUpdating files:  29% (3893/13423)\rUpdating files:  30% (4027/13423)\rUpdating files:  31% (4162/13423)\rUpdating files:  32% (4296/13423)\rUpdating files:  33% (4430/13423)\rUpdating files:  33% (4486/13423)\rUpdating files:  34% (4564/13423)\rUpdating files:  35% (4699/13423)\rUpdating files:  36% (4833/13423)\rUpdating files:  37% (4967/13423)\rUpdating files:  38% (5101/13423)\rUpdating files:  39% (5235/13423)\rUpdating files:  40% (5370/13423)\rUpdating files:  41% (5504/13423)\rUpdating files:  42% (5638/13423)\rUpdating files:  43% (5772/13423)\rUpdating files:  44% (5907/13423)\rUpdating files:  45% (6041/13423)\rUpdating files:  45% (6157/13423)\rUpdating files:  46% (6175/13423)\rUpdating files:  47% (6309/13423)\rUpdating files:  48% (6444/13423)\rUpdating files:  49% (6578/13423)\rUpdating files:  50% (6712/13423)\rUpdating files:  51% (6846/13423)\rUpdating files:  52% (6980/13423)\rUpdating files:  53% (7115/13423)\rUpdating files:  54% (7249/13423)\rUpdating files:  55% (7383/13423)\rUpdating files:  56% (7517/13423)\rUpdating files:  57% (7652/13423)\rUpdating files:  58% (7786/13423)\rUpdating files:  59% (7920/13423)\rUpdating files:  60% (8054/13423)\rUpdating files:  61% (8189/13423)\rUpdating files:  62% (8323/13423)\rUpdating files:  63% (8457/13423)\rUpdating files:  64% (8591/13423)\rUpdating files:  65% (8725/13423)\rUpdating files:  66% (8860/13423)\rUpdating files:  67% (8994/13423)\rUpdating files:  68% (9128/13423)\rUpdating files:  69% (9262/13423)\rUpdating files:  70% (9397/13423)\rUpdating files:  71% (9531/13423)\rUpdating files:  72% (9665/13423)\rUpdating files:  72% (9787/13423)\rUpdating files:  73% (9799/13423)\rUpdating files:  74% (9934/13423)\rUpdating files:  75% (10068/13423)\rUpdating files:  76% (10202/13423)\rUpdating files:  77% (10336/13423)\rUpdating files:  78% (10470/13423)\rUpdating files:  79% (10605/13423)\rUpdating files:  80% (10739/13423)\rUpdating files:  81% (10873/13423)\rUpdating files:  82% (11007/13423)\rUpdating files:  83% (11142/13423)\rUpdating files:  84% (11276/13423)\rUpdating files:  85% (11410/13423)\rUpdating files:  86% (11544/13423)\rUpdating files:  87% (11679/13423)\rUpdating files:  88% (11813/13423)\rUpdating files:  88% (11820/13423)\rUpdating files:  89% (11947/13423)\rUpdating files:  90% (12081/13423)\rUpdating files:  91% (12215/13423)\rUpdating files:  92% (12350/13423)\rUpdating files:  93% (12484/13423)\rUpdating files:  94% (12618/13423)\rUpdating files:  95% (12752/13423)\rUpdating files:  95% (12839/13423)\rUpdating files:  96% (12887/13423)\rUpdating files:  97% (13021/13423)\rUpdating files:  98% (13155/13423)\rUpdating files:  99% (13289/13423)\rUpdating files: 100% (13423/13423)\rUpdating files: 100% (13423/13423), done.\n        cluster.go:125: Note: switching to 'stable-4459.2.4-nightly-20260416-2100'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at 14b6d8f376 New version: stable-4459.2.4-nightly-20260416-2100\n        cluster.go:125: + emerge --getbinpkg --verbose coreos-sources\n        cluster.go:125: + zcat /proc/config.gz\n        cluster.go:125: ++ nproc\n        cluster.go:125: + exec make -C /usr/src/linux -j4 modules_prepare V=1\n        cluster.go:125: \n        cluster.go:125: Error in reading or end of file.\n"},{"name":"coreos.ignition.resource.local","result":"PASS","duration":54446061937,"output":""},{"name":"kubeadm.v1.33.8.flannel.base/node_readiness","result":"PASS","duration":10541564654,"output":""},{"name":"kubeadm.v1.33.8.flannel.base/nginx_deployment","result":"PASS","duration":10730846178,"output":""},{"name":"cl.flannel.udp","result":"PASS","duration":134394493109,"output":"        cluster.go:125: Timeout occurred while waiting for network connectivity.\n        flannel.go:121: ping from efad1e63-2d8d-4113-a28d-7f6cd0827b31(10.254.42.0) to 88ac6c8b-9de4-42d9-9425-d4a1caf0a737(10.254.34.1)\n"},{"name":"kubeadm.v1.33.8.flannel.base/NFS_deployment","result":"PASS","duration":17087692467,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.33.8.flannel.base","result":"PASS","duration":134823503340,"output":"        cluster.go:125: I0417 02:39:11.629130    1966 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0417 02:39:19.001041    2185 version.go:261] remote version is much newer: v1.35.4; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.147]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 502.28094ms\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.147:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.50843136s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 1.679514701s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 3.50171962s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: xad7od.zozlq0w4hssb1wx1\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.147:6443 --token xad7od.zozlq0w4hssb1wx1 \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:fdd8c2cdafab9285feb4383aa96b6cc016a08691c887958cfc6361dc8e218272 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        cluster.go:125: W0417 02:39:52.985188    1738 joinconfiguration.go:113] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.33.8.flannel.base/node_readiness (10.54s)\n    --- PASS: kubeadm.v1.33.8.flannel.base/nginx_deployment (10.73s)\n    --- PASS: kubeadm.v1.33.8.flannel.base/NFS_deployment (17.09s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:122): Cannot iterate over null (null)\n"}],"result":"FAIL","platform":"qemu","version":"4459.2.4"}
