{"tests":[{"name":"coreos.auth.verify","result":"PASS","duration":25114905003,"output":""},{"name":"cl.omaha.ping","result":"PASS","duration":25747199376,"output":""},{"name":"cl.ignition.v2_1.swap","result":"PASS","duration":25767299061,"output":""},{"name":"cl.ignition.v1.xfsroot","result":"PASS","duration":25771214763,"output":""},{"name":"cl.ignition.oem.indirect.new","result":"PASS","duration":50599961768,"output":""},{"name":"cl.disk.raid1.data","result":"PASS","duration":52494070379,"output":""},{"name":"coreos.ignition.resource.local","result":"PASS","duration":66754578192,"output":""},{"name":"cl.users.shells","result":"PASS","duration":52738854853,"output":""},{"name":"cl.ignition.v1.once","result":"FAIL","duration":245055031562,"output":"        harness.go:646: Cluster failed starting machines: machine \"9af9daee-c508-4b26-95c8-e8e925e9f0d4\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 9af9daee-c508-4b26-95c8-e8e925e9f0d4 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 9af9daee-c508-4b26-95c8-e8e925e9f0d4 console\n"},{"name":"cl.tpm.root-cryptenroll-pcr-noupdate","result":"PASS","duration":415204980549,"output":""},{"name":"cl.network.initramfs.second-boot","result":"PASS","duration":96911116356,"output":""},{"name":"docker.enable-service.sysext","result":"PASS","duration":62334712312,"output":"        cluster.go:152: + systemctl is-enabled docker\n"},{"name":"cl.cloudinit.multipart-mime","result":"FAIL","duration":627305700173,"output":"        harness.go:646: Cluster failed starting machines: machine \"c4b194b4-ed63-469c-869f-ad539365e0c1\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine c4b194b4-ed63-469c-869f-ad539365e0c1 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine c4b194b4-ed63-469c-869f-ad539365e0c1 console\n"},{"name":"cl.sysext.fallbackdownload","result":"FAIL","duration":604981217719,"output":"        update.go:633: creating test machine: machine \"8da5a8c7-c0f4-4c8b-abba-4f16e0c3f1c0\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.20:22: connect: no route to host\n"},{"name":"cl.ignition.translation","result":"FAIL","duration":824546782659,"output":"        harness.go:646: Cluster failed starting machines: machine \"9f560477-4b92-4ee6-8b6a-879f658aa21e\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 9f560477-4b92-4ee6-8b6a-879f658aa21e console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 9f560477-4b92-4ee6-8b6a-879f658aa21e console\n"},{"name":"linux.nfs.v4","result":"FAIL","duration":865840498241,"output":"        nfs.go:77: Cluster.NewMachine: machine \"6e97483b-36e3-48c1-8608-8684a68dc7e0\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 6e97483b-36e3-48c1-8608-8684a68dc7e0 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 6e97483b-36e3-48c1-8608-8684a68dc7e0 console\n"},{"name":"cl.ignition.symlink","result":"FAIL","duration":607583338068,"output":"        harness.go:646: Cluster failed starting machines: machine \"018c7f04-1576-427c-b2e7-8dfac6cfc47b\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.21:22: connect: no route to host\n"},{"name":"cl.internet/UpdateEngine","result":"PASS","duration":17126898944,"output":""},{"name":"cl.internet/DockerPing","result":"FAIL","duration":72097294937,"output":"        cluster.go:85: kolet: Process exited with status 1\n"},{"name":"cl.internet/DockerEcho","result":"FAIL","duration":21179194140,"output":"        cluster.go:85: kolet: Process exited with status 1\n"},{"name":"cl.internet/NTPDate","result":"PASS","duration":28495588267,"output":""},{"name":"cl.internet","result":"FAIL","duration":313395002838,"output":"    --- PASS: cl.internet/UpdateEngine (17.13s)\n        cluster.go:82: kolet:\n2026-04-13T20:20:41Z kolet: DockerPing timed out after 1m0s.\n    --- FAIL: cl.internet/DockerPing (72.10s)\n            cluster.go:85: kolet: Process exited with status 1\n        cluster.go:82: kolet:\n2026-04-13T20:21:02Z kolet: DockerEcho: exit status 125\n    --- FAIL: cl.internet/DockerEcho (21.18s)\n            cluster.go:85: kolet: Process exited with status 1\n    --- PASS: cl.internet/NTPDate (28.50s)\n"},{"name":"cl.ignition.instantiated.enable-unit","result":"FAIL","duration":1272642195873,"output":"        harness.go:646: Cluster failed starting machines: machine \"f91c5f6e-6836-4d3a-b32c-84f330f5464f\" failed basic checks: some systemd units failed:\n● session-2.scope             loaded failed failed Session 2 of User core\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine f91c5f6e-6836-4d3a-b32c-84f330f5464f console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine f91c5f6e-6836-4d3a-b32c-84f330f5464f console\n"},{"name":"docker.selinux","result":"FAIL","duration":621197337742,"output":"        harness.go:646: Cluster failed starting machines: machine \"5aee7bce-7c25-4cf3-97c7-0fd2ad595367\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.26:22: connect: no route to host\n"},{"name":"cl.cgroupv1","result":"FAIL","duration":626433806595,"output":"        harness.go:646: Cluster failed starting machines: machine \"de5c8512-4fee-4bf9-9237-92096a3ec13c\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.25:22: connect: no route to host\n"},{"name":"kubeadm.v1.33.8.calico.cgroupv1.base","result":"FAIL","duration":769606152174,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node with large disk: machine \"82b463ac-d5b7-4517-9142-15e04108c973\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.27:22: connect: no route to host\n"},{"name":"coreos.update.badusr","result":"FAIL","duration":616946436740,"output":"        harness.go:646: Cluster failed starting machines: machine \"3d74ca73-78c2-42d3-a702-c01c810a2960\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.29:22: connect: no route to host\n"},{"name":"cl.ignition.v1.groups","result":"FAIL","duration":611040003435,"output":"        harness.go:646: Cluster failed starting machines: machine \"44010233-3897-4207-89fc-3433555f31fa\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.30:22: connect: no route to host\n"},{"name":"kubeadm.v1.33.8.cilium.cgroupv1.base","result":"FAIL","duration":422100808631,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"690dbf45-610f-4b8f-a788-b442dac610cf\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 690dbf45-610f-4b8f-a788-b442dac610cf console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 690dbf45-610f-4b8f-a788-b442dac610cf console\n"},{"name":"cl.ignition.v1.users","result":"FAIL","duration":619664610582,"output":"        harness.go:646: Cluster failed starting machines: machine \"eee423be-0eee-4007-ba41-cc10c20d6e2f\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.33:22: connect: no route to host\n"},{"name":"docker.lib-coreos-dockerd-compat","result":"FAIL","duration":630744172526,"output":"        harness.go:646: Cluster failed starting machines: machine \"041d8fd5-130f-4c5d-bf47-c54750a12197\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.34:22: connect: no route to host\n"},{"name":"kubeadm.v1.33.8.cilium.base","result":"FAIL","duration":702428526477,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node with large disk: machine \"3b33a814-f98b-470b-b6a8-2c95b78c70ae\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.36:22: connect: no route to host\n"},{"name":"cl.tang.root","result":"FAIL","duration":461880766040,"output":"        tang.go:148: Started tang on 10.0.0.1:45205\n        tang.go:168: machine \"6b88a2e6-e282-4e17-b767-8b1d8e74f66a\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 6b88a2e6-e282-4e17-b767-8b1d8e74f66a console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-cryp…\u001b[0myptography Setup for rootencrypted.\r\r) on machine 6b88a2e6-e282-4e17-b767-8b1d8e74f66a console\n"},{"name":"cl.ignition.v2.xfsroot","result":"FAIL","duration":609769373003,"output":"        harness.go:646: Cluster failed starting machines: machine \"45d4562b-b7f2-4aa3-a1c4-2ddcbd1167c8\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.38:22: connect: no route to host\n"},{"name":"cl.ignition.v1.sethostname","result":"FAIL","duration":838559118799,"output":"        harness.go:646: Cluster failed starting machines: machine \"9bc1c253-6db1-481d-80a6-e3a0e28f9da0\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 9bc1c253-6db1-481d-80a6-e3a0e28f9da0 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 9bc1c253-6db1-481d-80a6-e3a0e28f9da0 console\n"},{"name":"cl.overlay.cleanup","result":"FAIL","duration":991773107128,"output":"        harness.go:646: Cluster failed starting machines: machine \"60cc58bd-ad11-4de8-b7e9-cc51af4e94a8\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 60cc58bd-ad11-4de8-b7e9-cc51af4e94a8 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 60cc58bd-ad11-4de8-b7e9-cc51af4e94a8 console\n"},{"name":"cl.disk.raid0.data","result":"FAIL","duration":459449864126,"output":"        raid.go:245: could not reboot machine: machine \"558768c7-f8d6-437d-8b2f-2ece636c65d0\" failed basic checks: systemctl: : Process exited with status 1: Failed to list units: Connection timed out\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 558768c7-f8d6-437d-8b2f-2ece636c65d0 console\n"},{"name":"cl.network.iptables","result":"FAIL","duration":601713423379,"output":"        harness.go:646: Cluster failed starting machines: machine \"f7424832-6b36-417e-8d88-6c7c2a3f0111\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.41:22: connect: no route to host\n"},{"name":"cl.etcd-member.discovery","result":"FAIL","duration":616383088497,"output":"        harness.go:646: Cluster failed starting machines: machine \"012cbec4-a0dd-4030-bfaf-f37fe6672811\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 012cbec4-a0dd-4030-bfaf-f37fe6672811 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 012cbec4-a0dd-4030-bfaf-f37fe6672811 console\n"},{"name":"coreos.selinux.boolean","result":"PASS","duration":138610094173,"output":""},{"name":"coreos.ignition.resource.remote","result":"PASS","duration":74875391671,"output":""},{"name":"systemd.journal.user","result":"FAIL","duration":858351165956,"output":"        harness.go:646: Cluster failed starting machines: machine \"89f50e63-eb55-4160-923a-bd5929e1173f\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 89f50e63-eb55-4160-923a-bd5929e1173f console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 89f50e63-eb55-4160-923a-bd5929e1173f console\n"},{"name":"coreos.selinux.enforce","result":"FAIL","duration":624878826768,"output":"        harness.go:646: Cluster failed starting machines: machine \"36d436ea-25f9-49b9-aabd-c7824280dd80\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.48:22: connect: connection refused\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 36d436ea-25f9-49b9-aabd-c7824280dd80 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 36d436ea-25f9-49b9-aabd-c7824280dd80 console\n"},{"name":"kubeadm.v1.35.1.calico.base","result":"FAIL","duration":621882754201,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"62e825e2-caee-4022-8896-610b5e3e0c2c\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.50:22: connect: no route to host\n"},{"name":"cl.etcd-member.v2-backup-restore","result":"FAIL","duration":1295855697088,"output":"        cluster.go:125: 2026-04-13T20:50:29Z\tinfo\tetcdutl/backup_command.go:216\tignoring EntryConfChange raft entry\n        cluster.go:125: 2026-04-13T20:50:29Z\tinfo\tetcdutl/backup_command.go:231\tignoring member attribute update on\t{\"entry\": \"Term:2 Index:3 Data:\\\"\\\\010\\\\202\\\\356\\\\226\\\\334\\\\211\\\\261\\\\247\\\\314\\\\352\\\\001\\\\022\\\\003PUT\\\\032\u0026/0/members/6aa7c0535a05ea98/attributes\\\\\\\"R{\\\\\\\"name\\\\\\\":\\\\\\\"3d1c737620244fce8b53121397b6fad4\\\\\\\",\\\\\\\"clientURLs\\\\\\\":[\\\\\\\"http://10.0.0.40:2379\\\\\\\"]}(\\\\0002\\\\0008\\\\000H\\\\000P\\\\000X\\\\000`\\\\000h\\\\000p\\\\000x\\\\000\\\\200\\\\001\\\\000\\\" \", \"v2Req.Path\": \"/0/members/6aa7c0535a05ea98/attributes\"}\n        cluster.go:125: 2026-04-13T20:50:29Z\tinfo\tetcdutl/backup_command.go:252\tignoring v3 raft entry\n        cluster.go:125: 2026-04-13T20:50:29Z\tinfo\tmembership/store.go:119\tTrimming membership information from the backend...\n        cluster.go:125: Reload daemon failed: Connection timed out\n        cluster.go:145: \"\\n\\tset -e\\n\\n\\texport ETCDCTL_API=2\\n\\tprefix=$RANDOM\\n\\tetcdctl set /$prefix/test magic\\n\\tres=\\\"$(etcdctl get /$prefix/test)\\\"\\n\\tif [[ \\\"$res\\\" != \\\"magic\\\" ]]; then\\n\\t\\techo \\\"Expected magic, got $res\\\"\\n\\t\\texit 1\\n\\tfi\\n\\n\\tbackup_to=\\\"$(mktemp -d)\\\"\\n\\n\\tsudo --preserve-env=ETCDCTL_API etcdctl backup --data-dir=/var/lib/etcd \\\\\\n\\t               --backup-dir \\\"${backup_to}\\\"\\n\\t\\n\\tetcdctl rm /$prefix/test\\n\\n\\tif etcdctl get /$prefix/test 2\u003e\u00261; then\\n\\t\\techo \\\"Expected rm'd key to error on get, didn't\\\"\\n\\t\\texit 1\\n\\tfi\\n\\n\\tsudo systemctl stop etcd-member\\n\\n\\t# Note: this means we're now a new cluster of size 1 because of how etcd2\\n\\t# backup/restore works.\\n\\tsudo rm -rf /var/lib/etcd\\n\\tsudo mv \\\"${backup_to}\\\" /var/lib/etcd/\\n\\tsudo chown -R etcd:etcd /var/lib/etcd\\n\\n\\tsudo mkdir -p /run/systemd/system/etcd-member.service.d/\\n\\tsudo tee /run/systemd/system/etcd-member.service.d/10-force-new.conf \u003c\u003cEOF\\n[Service]\\nEnvironment=ETCD_FORCE_NEW_CLUSTER=true\\nEOF\\n\\n\\tsudo systemctl daemon-reload\\n\\tsudo systemctl start etcd-member\\n\\n\\tres=\\\"$(etcdctl get /$prefix/test)\\\"\\n\\tif [[ \\\"$res\\\" != \\\"magic\\\" ]]; then\\n\\t\\techo \\\"Expected magic after backup-restore, got $res\\\"\\n\\t\\texit 1\\n\\tfi\\n\" failed: output magic\nPrevNode.Value: magic\nError:  100: Key not found (/6726/test) [5]\n[Service]\nEnvironment=ETCD_FORCE_NEW_CLUSTER=true, status Process exited with status 1\n"},{"name":"systemd.sysusers.gshadow","result":"FAIL","duration":552373504945,"output":"        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 5f4c5412-1675-4663-89d6-b5996ed19e0b console\n"},{"name":"coreos.ignition.once","result":"PASS","duration":815985191241,"output":""},{"name":"cl.ignition.oem.reuse","result":"PASS","duration":461081734855,"output":""},{"name":"cl.tpm.root","result":"FAIL","duration":622283262647,"output":"        tpm.go:324: machine \"e153d460-da9d-4849-905f-a702bb7cbbf0\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.55:22: connect: no route to host\n"},{"name":"cl.ignition.oem.indirect","result":"FAIL","duration":223786611600,"output":"        harness.go:646: Cluster failed starting machines: machine \"3bcf498a-4c69-4f8e-95a9-c3b652a4443a\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 3bcf498a-4c69-4f8e-95a9-c3b652a4443a console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 3bcf498a-4c69-4f8e-95a9-c3b652a4443a console\n"},{"name":"cl.network.wireguard","result":"FAIL","duration":1050711329472,"output":"        harness.go:646: Cluster failed starting machines: machine \"9d50f306-d397-4363-b68d-37435c217240\" failed basic checks: some systemd units failed:\n● session-2.scope  loaded failed failed Session 2 of User core\n● user@500.service loaded failed failed User Manager for UID 500\nstatus: \njournal:-- No entries --\n"},{"name":"cl.flannel.udp","result":"FAIL","duration":4126475876567,"output":"        harness.go:646: Cluster failed starting machines: machine \"ee3a4fea-00fb-4847-8bc5-871907291610\" failed basic checks: some systemd units failed:\n● flannel-docker-opts.service loaded failed failed start flannel docker export service - Network fabric for containers (System Application Container)\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 67cbcf56-3359-49d0-adaa-2360b05fdfce console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 67cbcf56-3359-49d0-adaa-2360b05fdfce console\n"},{"name":"cl.cloudinit.basic","result":"FAIL","duration":281501419066,"output":"        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39muser-configvirtfs.…\u001b[0mud-config from /media/configvirtfs.\r\r) on machine 5c332fd4-a2bb-444f-8c80-51485fccd30c console\n"},{"name":"coreos.ignition.systemd.enable-service","result":"PASS","duration":73270268129,"output":""},{"name":"docker.btrfs-storage","result":"PASS","duration":1110729311441,"output":""},{"name":"kubeadm.v1.34.4.cilium.base","result":"FAIL","duration":633317670531,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"7afb6ac0-c7eb-4cad-a5ad-0b52443f4266\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.61:22: connect: no route to host\n"},{"name":"cl.ignition.oem.wipe","result":"PASS","duration":229919835918,"output":""},{"name":"docker.containerd-restart","result":"FAIL","duration":1422505565780,"output":"        harness.go:646: Cluster failed starting machines: machine \"98071250-70b2-4290-af9c-d0e71faa4d32\" failed basic checks: some systemd units failed:\n● session-2.scope             loaded failed failed Session 2 of User core\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 98071250-70b2-4290-af9c-d0e71faa4d32 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 98071250-70b2-4290-af9c-d0e71faa4d32 console\n"},{"name":"coreos.locksmith.tls","result":"PASS","duration":1082673371057,"output":""},{"name":"kubeadm.v1.35.1.cilium.base","result":"FAIL","duration":692745416257,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"010faa3f-27ee-4181-a7f3-2bd9a878c637\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 010faa3f-27ee-4181-a7f3-2bd9a878c637 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 010faa3f-27ee-4181-a7f3-2bd9a878c637 console\n"},{"name":"linux.ntp","result":"PASS","duration":159184828206,"output":""},{"name":"sysext.custom-oem","result":"FAIL","duration":2581570744583,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-5951\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-5951\n        cluster.go:125: + sudo systemd-nspawn --console=pipe --bind-ro=/home/core/dev-container-script.sh --bind=/home/core/dev-container-workdir-5951:/work --image=flatcar_developer_container.bin --machine=flatcar-developer-container /bin/bash /home/core/dev-container-script.sh\n        cluster.go:125: + set -euo pipefail\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-13-1903\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-13-1903\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version_id=4081.3.7\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64-usr\n        cluster.go:125: + board=amd64-usr\n        cluster.go:125: + mkdir -p /work/sysext_rootfs/usr/share/flatcar-sysext-kola-test\n        cluster.go:125: + echo 4081.3.7\n        cluster.go:125: + mkdir -p /work/sysext_rootfs/usr/lib/extension-release.d\n        cluster.go:125: + sysext_arch=x86-64\n        cluster.go:125: + [[ amd64-usr = \\a\\r\\m\\6\\4\\-\\u\\s\\r ]]\n        cluster.go:125: + metadata=('ID=flatcar' \"VERSION_ID=${version_id}\" \"ARCHITECTURE=${sysext_arch}\")\n        cluster.go:125: + metadata_file=/work/sysext_rootfs/usr/lib/extension-release.d/extension-release.oem-test\n        cluster.go:125: + printf '%s\\n' ID=flatcar VERSION_ID=4081.3.7 ARCHITECTURE=x86-64\n        cluster.go:125: + mksquashfs /work/sysext_rootfs /work/oem-test-4081.3.7.raw -all-root\n        cluster.go:125: ++ source /etc/os-release\n        cluster.go:125: +++ NAME='Flatcar Container Linux by Kinvolk'\n        cluster.go:125: +++ ID=flatcar\n        cluster.go:125: +++ ID_LIKE=coreos\n        cluster.go:125: +++ VERSION=4081.3.7\n        cluster.go:125: +++ VERSION_ID=4081.3.7\n        cluster.go:125: +++ BUILD_ID=2026-04-13-1915\n        cluster.go:125: +++ SYSEXT_LEVEL=1.0\n        cluster.go:125: +++ PRETTY_NAME='Flatcar Container Linux by Kinvolk 4081.3.7 (LTS 2024)'\n        cluster.go:125: +++ ANSI_COLOR='38;5;75'\n        cluster.go:125: +++ HOME_URL=https://flatcar.org/\n        cluster.go:125: +++ BUG_REPORT_URL=https://issues.flatcar.org\n        cluster.go:125: +++ FLATCAR_BOARD=amd64-usr\n        cluster.go:125: +++ CPE_NAME='cpe:2.3:o:flatcar-linux:flatcar_linux:4081.3.7:*:*:*:*:*:*:*'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: + sysext_file=/home/core/dev-container-workdir-5951/oem-test-4081.3.7.raw\n        cluster.go:125: + [[ ! -e /home/core/dev-container-workdir-5951/oem-test-4081.3.7.raw ]]\n        cluster.go:125: + [[ ! -e /oem/oem-release ]]\n        cluster.go:125: + printf '%s\\n' ID=test VERSION_ID=1.0.0 'NAME=testing stuff'\n        cluster.go:125: + sudo tee /oem/oem-release\n        cluster.go:125: + sudo mkdir -p /oem/sysext\n        cluster.go:125: + sudo mv /home/core/dev-container-workdir-5951/oem-test-4081.3.7.raw /oem/sysext\n        cluster.go:125: + sudo touch /oem/sysext/active-oem-test\n        cluster.go:125: + sudo flatcar-reset --keep-machine-id --keep-paths /var/log\n        sysext.go:394: could not reboot: machine \"791b29a0-96dc-49f8-9a7a-4e706e7cdeeb\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 791b29a0-96dc-49f8-9a7a-4e706e7cdeeb console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 791b29a0-96dc-49f8-9a7a-4e706e7cdeeb console\n"},{"name":"sysext.disable-docker","result":"PASS","duration":42988769964,"output":""},{"name":"devcontainer.docker","result":"FAIL","duration":629589426425,"output":"        devcontainer.go:228: creating a machine failed: machine \"68783d6f-7d10-4bad-b365-8d2b35eca3c2\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.65:22: connect: no route to host\n"},{"name":"coreos.tls.fetch-urls","result":"FAIL","duration":387108358758,"output":"        harness.go:646: Cluster failed starting machines: machine \"426cca75-bb0a-4b56-ac9a-27258008b9d4\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 426cca75-bb0a-4b56-ac9a-27258008b9d4 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 426cca75-bb0a-4b56-ac9a-27258008b9d4 console\n"},{"name":"cl.swap_activation","result":"PASS","duration":657081358236,"output":""},{"name":"bpf.ig/ig","result":"FAIL","duration":468224870845,"output":"        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n        cluster.go:125: time=\"2026-04-13T21:25:39Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Get \\\"http://%2Frun%2Fdocker.sock/v1.51/containers/json?all=1\\\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\"\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n        cluster.go:125: ++ kill %%\n        ig.go:79: ig run trace_exec did not behave as expected: Process exited with status 124\n"},{"name":"bpf.ig","result":"FAIL","duration":470504863457,"output":"    --- FAIL: bpf.ig/ig (468.22s)\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n            cluster.go:125: time=\"2026-04-13T21:25:39Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Get \\\"http://%2Frun%2Fdocker.sock/v1.51/containers/json?all=1\\\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\"\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n            cluster.go:125: ++ kill %%\n            ig.go:79: ig run trace_exec did not behave as expected: Process exited with status 124\n"},{"name":"cl.update.badverity","result":"PASS","duration":918188158380,"output":""},{"name":"coreos.ignition.groups","result":"PASS","duration":51467450419,"output":""},{"name":"cl.cloudinit.script","result":"PASS","duration":49550733682,"output":""},{"name":"cl.tang.nonroot","result":"PASS","duration":797707303860,"output":"        tang.go:148: Started tang on 10.0.0.1:41317\n"},{"name":"cl.tpm.root-cryptenroll","result":"FAIL","duration":1196852701547,"output":"        tpm.go:341: could not reboot machine: machine \"946bdd2b-52e1-48e5-8fc6-9568b99c8434\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 946bdd2b-52e1-48e5-8fc6-9568b99c8434 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 946bdd2b-52e1-48e5-8fc6-9568b99c8434 console\n"},{"name":"linux.nfs.v3","result":"PASS","duration":840306706283,"output":"        nfs.go:80: NFS server booted.\n        nfs.go:85: Test file \"/tmp/tmp.0hrn1YGwPL\" created on server.\n        nfs.go:125: NFS client booted.\n        nfs.go:133: Got NFS mount.\n"},{"name":"cl.filesystem/deadlinks","result":"PASS","duration":141889292260,"output":""},{"name":"cl.filesystem/suid","result":"PASS","duration":9492891872,"output":""},{"name":"cl.ignition.v2_1.ext4checkexisting","result":"PASS","duration":876734348083,"output":""},{"name":"cl.filesystem/sgid","result":"PASS","duration":11211023973,"output":""},{"name":"cl.filesystem/writablefiles","result":"PASS","duration":23185012789,"output":""},{"name":"cl.filesystem/writabledirs","result":"PASS","duration":4343155336,"output":""},{"name":"cl.filesystem/stickydirs","result":"PASS","duration":12685602608,"output":""},{"name":"cl.filesystem/blacklist","result":"PASS","duration":9139801270,"output":""},{"name":"cl.filesystem","result":"PASS","duration":468255401768,"output":"    --- PASS: cl.filesystem/deadlinks (141.89s)\n    --- PASS: cl.filesystem/suid (9.49s)\n    --- PASS: cl.filesystem/sgid (11.21s)\n    --- PASS: cl.filesystem/writablefiles (23.19s)\n    --- PASS: cl.filesystem/writabledirs (4.34s)\n    --- PASS: cl.filesystem/stickydirs (12.69s)\n    --- PASS: cl.filesystem/blacklist (9.14s)\n"},{"name":"cl.tpm.root-cryptenroll-pcr-withupdate","result":"PASS","duration":550797247241,"output":"        cluster.go:125: New TPM2 token enrolled as key slot 1.\n        cluster.go:125: Wiped slot 2.\n"},{"name":"kubeadm.v1.33.8.flannel.cgroupv1.base","result":"FAIL","duration":684887046811,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node with large disk: machine \"360f13d2-a685-4a70-bba2-c946732f4d12\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.86:22: connect: connection refused\n"},{"name":"coreos.locksmith.reboot","result":"FAIL","duration":1202551577628,"output":"        locksmith.go:141: failed to check rebooted machine: ssh unreachable or system not ready: context deadline exceeded\n"},{"name":"cl.disk.raid1.root","result":"PASS","duration":613383199563,"output":""},{"name":"cl.ignition.oem.regular.new","result":"PASS","duration":108958672828,"output":""},{"name":"cl.basic","result":"FAIL","duration":313431379698,"output":"        harness.go:646: Cluster failed starting machines: machine \"0a1fb6c8-3a91-40e6-ba3a-1442adead007\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 0a1fb6c8-3a91-40e6-ba3a-1442adead007 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 0a1fb6c8-3a91-40e6-ba3a-1442adead007 console\n"},{"name":"cl.ignition.luks","result":"FAIL","duration":597203359954,"output":"        harness.go:646: Cluster failed starting machines: machine \"4318aa1b-ed7c-4cfc-872b-f8b9fe2fdae7\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 4318aa1b-ed7c-4cfc-872b-f8b9fe2fdae7 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 4318aa1b-ed7c-4cfc-872b-f8b9fe2fdae7 console\n"},{"name":"packages","result":"FAIL","duration":786656355875,"output":"        harness.go:646: Cluster failed starting machines: machine \"2c9e9c9c-01b9-42af-9da3-e040ccc82ee4\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 2c9e9c9c-01b9-42af-9da3-e040ccc82ee4 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 2c9e9c9c-01b9-42af-9da3-e040ccc82ee4 console\n"},{"name":"cl.toolbox.dnf-install","result":"PASS","duration":1293534771715,"output":""},{"name":"cl.ignition.v1.btrfsroot","result":"PASS","duration":417395339247,"output":""},{"name":"sysext.simple","result":"PASS","duration":175773373687,"output":""},{"name":"docker.network-openbsd-nc","result":"FAIL","duration":1775792463309,"output":"        docker.go:413: creating netcat containers\n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 1.1s done\n        cluster.go:125: #1 DONE 5.5s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B 1.0s done\n        cluster.go:125: #2 DONE 4.6s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 0.1s\n        cluster.go:125: #3 transferring context: 1.80MB 5.4s\n        cluster.go:125: #3 transferring context: 2.45MB 7.1s done\n        cluster.go:125: #3 DONE 11.5s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 9.9s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 9.5s done\n        cluster.go:125: #5 writing image sha256:adfce8466c7bbfd3fdea7f10e27729d9e5f7fd40a2cbf666c49ce088e12b3393\n        cluster.go:125: #5 writing image sha256:adfce8466c7bbfd3fdea7f10e27729d9e5f7fd40a2cbf666c49ce088e12b3393 0.8s done\n        cluster.go:125: #5 naming to docker.io/library/netcat\n        cluster.go:125: #5 naming to docker.io/library/netcat 2.2s done\n        cluster.go:125: #5 DONE 17.8s\n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 1.0s done\n        cluster.go:125: #1 DONE 2.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B 0.7s done\n        cluster.go:125: #2 DONE 3.6s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context:\n        cluster.go:125: #3 transferring context: 1.17MB 5.1s\n        cluster.go:125: #3 transferring context: 1.93MB 10.4s\n        cluster.go:125: #3 transferring context: 2.45MB 13.8s done\n        cluster.go:125: #3 DONE 17.2s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 1.6s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 1.0s done\n        cluster.go:125: #5 writing image sha256:97df076f63e10e422a023ef26474fcfcbf32d92e42044918f95411f491441bef\n        cluster.go:125: #5 writing image sha256:97df076f63e10e422a023ef26474fcfcbf32d92e42044918f95411f491441bef 0.3s done\n        cluster.go:125: #5 naming to docker.io/library/netcat\n        cluster.go:125: #5 naming to docker.io/library/netcat 0.6s done\n        cluster.go:125: #5 DONE 3.5s\n        docker.go:471: [0] Process exited with status 124 [1] Process exited with status 1\n"},{"name":"cl.ignition.v2.users","result":"PASS","duration":100778568024,"output":""},{"name":"kubeadm.v1.35.1.flannel.base","result":"FAIL","duration":1802526588043,"output":"        kubeadm.go:197: unable to setup cluster: unable to get etcd node health: health polling failed: Process exited with status 1: \n"},{"name":"docker.base","result":"FAIL","duration":626496458862,"output":"        harness.go:646: Cluster failed starting machines: machine \"82f27821-5c90-41a1-8111-8732340ec0de\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.102:22: connect: no route to host\n"},{"name":"coreos.ignition.resource.s3.versioned","result":"FAIL","duration":1391704789960,"output":"        harness.go:646: Cluster failed starting machines: machine \"836a1558-74ff-48fd-9a56-ee48eea3d0b4\" failed basic checks: some systemd units failed:\n● session-3.scope  loaded failed failed Session 3 of User core\n● user@500.service loaded failed failed User Manager for UID 500\nstatus: \njournal:-- No entries --\n"},{"name":"cl.ignition.v2.ext4root","result":"PASS","duration":765043728078,"output":""},{"name":"cl.update.reboot","result":"FAIL","duration":740432092675,"output":"        harness.go:646: Cluster failed starting machines: machine \"d945d915-c084-49f1-a4c6-7d75d8db1581\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine d945d915-c084-49f1-a4c6-7d75d8db1581 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine d945d915-c084-49f1-a4c6-7d75d8db1581 console\n"},{"name":"kubeadm.v1.33.8.flannel.base","result":"FAIL","duration":627147173145,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"21ae9de1-461e-4cac-b6e9-5e82dd4aa3e1\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.109:22: connect: no route to host\n"},{"name":"misc.fips","result":"PASS","duration":345584279248,"output":"        cluster.go:125: Error setting digest\n        cluster.go:125: 4017E869D37F0000:error:0308010C:digital envelope routines:inner_evp_generic_fetch:unsupported:../openssl-3.2.3/crypto/evp/evp_fetch.c:355:Global default library context, Algorithm (MD5 : 102), Properties ()\n        cluster.go:125: 4017E869D37F0000:error:03000086:digital envelope routines:evp_md_init_internal:initialization error:../openssl-3.2.3/crypto/evp/digest.c:272:\n        cluster.go:152: + cat /proc/sys/crypto/fips_enabled\n"},{"name":"systemd.journal.remote","result":"FAIL","duration":907455413894,"output":"        journald.go:143: Cluster.NewMachine: machine \"9bc84b9a-fa28-4335-b569-7b52883ca757\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 9bc84b9a-fa28-4335-b569-7b52883ca757 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 9bc84b9a-fa28-4335-b569-7b52883ca757 console\n"},{"name":"cl.sysext.boot","result":"FAIL","duration":377304455843,"output":"        update.go:538: creating test machine: machine \"4f4c40a5-e1f2-4996-b805-355a68133aa7\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 4f4c40a5-e1f2-4996-b805-355a68133aa7 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 4f4c40a5-e1f2-4996-b805-355a68133aa7 console\n"},{"name":"cl.etcd-member.etcdctlv3","result":"PASS","duration":341534727993,"output":"        cluster.go:125: {\"level\":\"info\",\"ts\":1776119130.019029,\"caller\":\"snapshot/v3_snapshot.go:68\",\"msg\":\"created temporary db file\",\"path\":\"/tmp/tmp.Wzgdt1kwJB/snapshot.db.part\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776119130.0813227,\"logger\":\"client\",\"caller\":\"v3/maintenance.go:211\",\"msg\":\"opened snapshot stream; downloading\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776119130.0816615,\"caller\":\"snapshot/v3_snapshot.go:76\",\"msg\":\"fetching snapshot\",\"endpoint\":\"127.0.0.1:2379\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776119130.233669,\"logger\":\"client\",\"caller\":\"v3/maintenance.go:219\",\"msg\":\"completed snapshot read; closing\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776119130.2338505,\"caller\":\"snapshot/v3_snapshot.go:91\",\"msg\":\"fetched snapshot\",\"endpoint\":\"127.0.0.1:2379\",\"size\":\"20 kB\",\"took\":\"now\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":1776119130.2338936,\"caller\":\"snapshot/v3_snapshot.go:100\",\"msg\":\"saved\",\"path\":\"/tmp/tmp.Wzgdt1kwJB/snapshot.db\"}\n        cluster.go:125: Deprecated: Use `etcdutl snapshot status` instead.\n"},{"name":"cl.flannel.vxlan","result":"FAIL","duration":3009776670746,"output":"        harness.go:646: Cluster failed starting machines: machine \"6a14da50-f977-4d7d-90d1-8cab8337be7d\" failed basic checks: ssh unreachable or system not ready: failure checking if machine is running: systemctl is-system-running returned stdout: \"starting\", stderr: \"\", err: Process exited with status 1, systemctl list-jobs returned stdout: \"JOB   UNIT                        TYPE  STATE\\n25606 flanneld.service            start running\\n25605 flannel-docker-opts.service start waiting\\n476   multi-user.target           start waiting\\n\\n3 jobs listed.\", stderr: \"\", err: \u003cnil\u003e\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 6a14da50-f977-4d7d-90d1-8cab8337be7d console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 6a14da50-f977-4d7d-90d1-8cab8337be7d console\n"},{"name":"cl.ignition.partition_on_boot_disk","result":"FAIL","duration":614843858202,"output":"        filesystem.go:484: machine \"cbca17fb-4d9b-474e-8084-2093733aeda1\" failed to start: ssh journalctl failed: time limit exceeded: ssh: handshake failed: read tcp 10.0.0.1:46466-\u003e10.0.0.119:22: read: connection reset by peer\n        harness.go:616: Found emergency shell on machine cbca17fb-4d9b-474e-8084-2093733aeda1 console\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine cbca17fb-4d9b-474e-8084-2093733aeda1 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine cbca17fb-4d9b-474e-8084-2093733aeda1 console\n"},{"name":"sysext.disable-containerd","result":"FAIL","duration":996705609293,"output":"        harness.go:646: Cluster failed starting machines: machine \"8d514457-5cb5-4afe-b94f-924fb03d8fb0\" failed basic checks: some systemd units failed:\n● session-2.scope             loaded failed failed Session 2 of User core\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 8d514457-5cb5-4afe-b94f-924fb03d8fb0 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 8d514457-5cb5-4afe-b94f-924fb03d8fb0 console\n"},{"name":"cl.ignition.oem.regular","result":"FAIL","duration":1241341235805,"output":"        oem.go:199: Couldn't reboot machine: machine \"5bfb6b63-c3de-4a1a-8c4c-782b3005ed72\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.115:22: connect: connection refused\n"},{"name":"cl.ignition.v1.ext4root","result":"PASS","duration":297488673064,"output":""},{"name":"kubeadm.v1.34.4.flannel.base","result":"FAIL","duration":617784806955,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"2389c32d-b26d-47b2-a737-62b0c42e95cf\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.121:22: connect: no route to host\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 2389c32d-b26d-47b2-a737-62b0c42e95cf console\n"},{"name":"cl.ignition.kargs","result":"PASS","duration":374180794531,"output":"        cluster.go:152: + cat /proc/cmdline\n"},{"name":"coreos.ignition.sethostname","result":"PASS","duration":65306694934,"output":""},{"name":"docker.userns","result":"FAIL","duration":1517601372338,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.6s done\n        cluster.go:125: #1 DONE 1.9s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B 0.2s done\n        cluster.go:125: #2 DONE 1.6s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context:\n        cluster.go:125: #3 transferring context: 2.24MB 4.0s done\n        cluster.go:125: #3 DONE 6.5s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 3.9s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 1.2s done\n        cluster.go:125: #5 writing image sha256:e981511fc0349f4bd7f817fe49e3ad9ab2ae25c251da8e1c34fb63dff65e011d 0.0s done\n        cluster.go:125: #5 naming to docker.io/library/userns-test\n        cluster.go:125: #5 naming to docker.io/library/userns-test 1.2s done\n        cluster.go:125: #5 DONE 6.0s\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 93a58224-2bda-4997-a2c5-24e943cc9b3f console\n"},{"name":"cl.ignition.v2_1.vfat","result":"PASS","duration":108896908554,"output":""},{"name":"cl.locksmith.cluster","result":"FAIL","duration":2224932466133,"output":"        locksmith.go:184: [0] ssh unreachable or system not ready: context deadline exceeded [1] ssh unreachable or system not ready: context deadline exceeded [2] ssh unreachable or system not ready: context deadline exceeded\n"},{"name":"cl.disk.raid0.root","result":"PASS","duration":134035387011,"output":""},{"name":"cl.network.listeners","result":"PASS","duration":779920897765,"output":""},{"name":"cl.osreset.ignition-rerun","result":"PASS","duration":1169437481885,"output":""},{"name":"cl.misc.falco","result":"PASS","duration":3454279322303,"output":"        cluster.go:125: Unable to find image 'falcosecurity/falco-driver-loader:master' locally\n        cluster.go:125: master: Pulling from falcosecurity/falco-driver-loader\n        cluster.go:125: 0858e4838ecf: Pulling fs layer\n        cluster.go:125: 2b9604c8ac9b: Pulling fs layer\n        cluster.go:125: 9a34fcea6483: Pulling fs layer\n        cluster.go:125: 1f55ac5619ae: Pulling fs layer\n        cluster.go:125: 677364613154: Pulling fs layer\n        cluster.go:125: de07b784e5b1: Pulling fs layer\n        cluster.go:125: dac426886d53: Pulling fs layer\n        cluster.go:125: b670aed5aa2d: Pulling fs layer\n        cluster.go:125: 1f55ac5619ae: Waiting\n        cluster.go:125: 677364613154: Waiting\n        cluster.go:125: de07b784e5b1: Waiting\n        cluster.go:125: dac426886d53: Waiting\n        cluster.go:125: b670aed5aa2d: Waiting\n        cluster.go:125: 2b9604c8ac9b: Verifying Checksum\n        cluster.go:125: 2b9604c8ac9b: Download complete\n        cluster.go:125: 1f55ac5619ae: Verifying Checksum\n        cluster.go:125: 1f55ac5619ae: Download complete\n        cluster.go:125: 677364613154: Verifying Checksum\n        cluster.go:125: 677364613154: Download complete\n        cluster.go:125: 0858e4838ecf: Verifying Checksum\n        cluster.go:125: 0858e4838ecf: Download complete\n        cluster.go:125: dac426886d53: Verifying Checksum\n        cluster.go:125: dac426886d53: Download complete\n        cluster.go:125: b670aed5aa2d: Download complete\n        cluster.go:125: 9a34fcea6483: Download complete\n        cluster.go:125: de07b784e5b1: Verifying Checksum\n        cluster.go:125: de07b784e5b1: Download complete\n        cluster.go:125: 0858e4838ecf: Pull complete\n        cluster.go:125: 2b9604c8ac9b: Pull complete\n        cluster.go:125: 9a34fcea6483: Pull complete\n        cluster.go:125: 1f55ac5619ae: Pull complete\n        cluster.go:125: 677364613154: Pull complete\n        cluster.go:125: de07b784e5b1: Pull complete\n        cluster.go:125: dac426886d53: Pull complete\n        cluster.go:125: b670aed5aa2d: Pull complete\n        cluster.go:125: Digest: sha256:0871e09c1cab800ab96d9a7e3e25118e33f7bbc45fbc53563c7933847b189535\n        cluster.go:125: Status: Downloaded newer image for falcosecurity/falco-driver-loader:master\n"},{"name":"cl.verity/verify","result":"PASS","duration":285014867338,"output":"        cluster.go:125: Success\n"},{"name":"sysext.custom-docker.sysext","result":"PASS","duration":1499338347608,"output":"        cluster.go:125: bash: line 1: docker: command not found\n        cluster.go:125: Cloning into 'sysext-bakery'...\n        cluster.go:125: Updating files:  49% (93/189)\rUpdating files:  50% (95/189)\rUpdating files:  51% (97/189)\rUpdating files:  52% (99/189)\rUpdating files:  53% (101/189)\rUpdating files:  54% (103/189)\rUpdating files:  55% (104/189)\rUpdating files:  56% (106/189)\rUpdating files:  57% (108/189)\rUpdating files:  57% (109/189)\rUpdating files:  58% (110/189)\rUpdating files:  59% (112/189)\rUpdating files:  60% (114/189)\rUpdating files:  61% (116/189)\rUpdating files:  62% (118/189)\rUpdating files:  63% (120/189)\rUpdating files:  64% (121/189)\rUpdating files:  65% (123/189)\rUpdating files:  66% (125/189)\rUpdating files:  66% (126/189)\rUpdating files:  67% (127/189)\rUpdating files:  68% (129/189)\rUpdating files:  69% (131/189)\rUpdating files:  70% (133/189)\rUpdating files:  71% (135/189)\rUpdating files:  72% (137/189)\rUpdating files:  73% (138/189)\rUpdating files:  74% (140/189)\rUpdating files:  75% (142/189)\rUpdating files:  76% (144/189)\rUpdating files:  77% (146/189)\rUpdating files:  78% (148/189)\rUpdating files:  79% (150/189)\rUpdating files:  80% (152/189)\rUpdating files:  81% (154/189)\rUpdating files:  82% (155/189)\rUpdating files:  83% (157/189)\rUpdating files:  83% (158/189)\rUpdating files:  84% (159/189)\rUpdating files:  85% (161/189)\rUpdating files:  86% (163/189)\rUpdating files:  87% (165/189)\rUpdating files:  87% (166/189)\rUpdating files:  88% (167/189)\rUpdating files:  89% (169/189)\rUpdating files:  90% (171/189)\rUpdating files:  91% (172/189)\rUpdating files:  92% (174/189)\rUpdating files:  93% (176/189)\rUpdating files:  94% (178/189)\rUpdating files:  95% (180/189)\rUpdating files:  96% (182/189)\rUpdating files:  97% (184/189)\rUpdating files:  98% (186/189)\rUpdating files:  99% (188/189)\rUpdating files: 100% (189/189)\rUpdating files: 100% (189/189), done.\n        cluster.go:125: Note: switching to '9850ffd5b2353f45a9b3bf4fb84f8138a149e3e7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at 9850ffd Merge pull request #31 from flatcar/t-lo/fix-docker-23-containerd-shim\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Download complete\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n"},{"name":"cl.install.cloudinit","result":"PASS","duration":62993282860,"output":""},{"name":"coreos.ignition.security.tls","result":"PASS","duration":150040929008,"output":""},{"name":"cl.verity/corruption","result":"PASS","duration":185706940408,"output":""},{"name":"cl.verity","result":"PASS","duration":724845004244,"output":"    --- PASS: cl.verity/verify (285.01s)\n            cluster.go:125: Success\n    --- PASS: cl.verity/corruption (185.71s)\n"},{"name":"cl.tpm.nonroot","result":"PASS","duration":539504781684,"output":""},{"name":"cl.ignition.v2.btrfsroot","result":"PASS","duration":237467558983,"output":""},{"name":"kubeadm.v1.33.8.calico.base","result":"FAIL","duration":1776647135034,"output":"        cluster.go:125: I0413 23:00:39.822830    1864 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0413 23:09:06.939841    2331 version.go:261] remote version is much newer: v1.35.3; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.10\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 23:09:27.405279    2331 checks.go:843] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm.It is recommended to use \"registry.k8s.io/pause:3.10\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.131]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 16.568312552s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.131:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 19.127184871s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 40.702763795s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 1m33.168126386s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: d7m2vo.pgeuxpn1ykb9lzx0\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.131:6443 --token d7m2vo.pgeuxpn1ykb9lzx0 \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:f6a7c39f0b61d2e8a963beaa399d48b319b919a8e2d61cb86a160f48c3431636 \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: error: timed out waiting for the condition\n        kubeadm.go:197: unable to setup cluster: unable to run master script: Process exited with status 1\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 57b66e2c-9d54-4325-852e-81f66eac64ec console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 57b66e2c-9d54-4325-852e-81f66eac64ec console\n"},{"name":"kubeadm.v1.34.4.calico.base/node_readiness","result":"PASS","duration":124540448927,"output":""},{"name":"kubeadm.v1.34.4.calico.base/nginx_deployment","result":"PASS","duration":11923134030,"output":""},{"name":"kubeadm.v1.34.4.calico.base/NFS_deployment","result":"PASS","duration":19904932750,"output":"        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n        cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"kubeadm.v1.34.4.calico.base","result":"PASS","duration":2021109327418,"output":"        cluster.go:125: W0413 23:08:29.711002    1930 version.go:108] could not fetch a Kubernetes version from the internet: unable to get URL \"https://dl.k8s.io/release/stable-1.txt\": Get \"https://dl.k8s.io/release/stable-1.txt\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\n        cluster.go:125: W0413 23:08:29.711998    1930 version.go:109] falling back to the local client version: v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0413 23:14:16.644485    2352 version.go:260] remote version is much newer: v1.35.3; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.6\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 23:14:26.768303    2352 checks.go:827] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.138]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 15.563008029s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.138:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 11.140609892s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 35.798096816s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 4m1.592110166s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: 5f2dhb.9ivknk1q05jk6lw1\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.138:6443 --token 5f2dhb.9ivknk1q05jk6lw1 \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:5f985af3d9302034a1d225692643286e316c75088a9f17f46e9bfc94964488b9 \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: installation.operator.tigera.io/default created\n        cluster.go:125: apiserver.operator.tigera.io/default created\n        cluster.go:125: goldmane.operator.tigera.io/default created\n        cluster.go:125: whisker.operator.tigera.io/default created\n        cluster.go:125: W0413 23:29:49.495097    1647 joinconfiguration.go:112] [config] WARNING: Ignored configuration document with GroupVersionKind kubelet.config.k8s.io/v1beta1, Kind=KubeletConfiguration\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n    --- PASS: kubeadm.v1.34.4.calico.base/node_readiness (124.54s)\n    --- PASS: kubeadm.v1.34.4.calico.base/nginx_deployment (11.92s)\n    --- PASS: kubeadm.v1.34.4.calico.base/NFS_deployment (19.90s)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n            cluster.go:125: jq: error (at \u003cstdin\u003e:123): Cannot iterate over null (null)\n"},{"name":"devcontainer.systemd-nspawn","result":"PASS","duration":7709881495238,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + ARCH=amd64\n        cluster.go:125: + VERSION=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + EXPECTED_VERSION=4081.3.7\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-12516\n        cluster.go:125: + USR_SRC_DIR=/home/core/dev-container-workdir-12516/src\n        cluster.go:125: + VAR_TMP_DIR=/home/core/dev-container-workdir-12516/tmp\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-12516/src /home/core/dev-container-workdir-12516/tmp\n        cluster.go:125: + sudo systemd-nspawn --console=pipe --setenv=PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs --setenv=EXPECTED_VERSION=4081.3.7 --bind-ro=/lib/modules --bind-ro=/home/core/dev-container-script --bind=/home/core/dev-container-workdir-12516/src:/usr/src --bind=/home/core/dev-container-workdir-12516/tmp:/var/tmp --image=flatcar_developer_container.bin --machine=flatcar-developer-container /bin/bash /home/core/dev-container-script\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + [[ 4081.3.7 != \\4\\0\\8\\1\\.\\3\\.\\7 ]]\n        cluster.go:125: + export PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + export 'FEATURES=-ipc-sandbox -network-sandbox'\n        cluster.go:125: + FEATURES='-ipc-sandbox -network-sandbox'\n        cluster.go:125: + emerge-gitclone\n        cluster.go:125: Cloning into '/var/lib/portage/scripts'...\n        cluster.go:125: Updating files:  20% (2753/13424)\rUpdating files:  21% (2820/13424)\rUpdating files:  22% (2954/13424)\rUpdating files:  23% (3088/13424)\rUpdating files:  24% (3222/13424)\rUpdating files:  25% (3356/13424)\rUpdating files:  26% (3491/13424)\rUpdating files:  27% (3625/13424)\rUpdating files:  28% (3759/13424)\rUpdating files:  29% (3893/13424)\rUpdating files:  30% (4028/13424)\rUpdating files:  31% (4162/13424)\rUpdating files:  32% (4296/13424)\rUpdating files:  33% (4430/13424)\rUpdating files:  34% (4565/13424)\rUpdating files:  34% (4625/13424)\rUpdating files:  35% (4699/13424)\rUpdating files:  36% (4833/13424)\rUpdating files:  37% (4967/13424)\rUpdating files:  38% (5102/13424)\rUpdating files:  39% (5236/13424)\rUpdating files:  40% (5370/13424)\rUpdating files:  41% (5504/13424)\rUpdating files:  42% (5639/13424)\rUpdating files:  43% (5773/13424)\rUpdating files:  44% (5907/13424)\rUpdating files:  45% (6041/13424)\rUpdating files:  46% (6176/13424)\rUpdating files:  47% (6310/13424)\rUpdating files:  48% (6444/13424)\rUpdating files:  49% (6578/13424)\rUpdating files:  50% (6712/13424)\rUpdating files:  51% (6847/13424)\rUpdating files:  52% (6981/13424)\rUpdating files:  53% (7115/13424)\rUpdating files:  53% (7117/13424)\rUpdating files:  54% (7249/13424)\rUpdating files:  55% (7384/13424)\rUpdating files:  56% (7518/13424)\rUpdating files:  57% (7652/13424)\rUpdating files:  58% (7786/13424)\rUpdating files:  59% (7921/13424)\rUpdating files:  60% (8055/13424)\rUpdating files:  61% (8189/13424)\rUpdating files:  62% (8323/13424)\rUpdating files:  63% (8458/13424)\rUpdating files:  64% (8592/13424)\rUpdating files:  65% (8726/13424)\rUpdating files:  65% (8817/13424)\rUpdating files:  66% (8860/13424)\rUpdating files:  67% (8995/13424)\rUpdating files:  68% (9129/13424)\rUpdating files:  69% (9263/13424)\rUpdating files:  70% (9397/13424)\rUpdating files:  71% (9532/13424)\rUpdating files:  72% (9666/13424)\rUpdating files:  73% (9800/13424)\rUpdating files:  74% (9934/13424)\rUpdating files:  75% (10068/13424)\rUpdating files:  76% (10203/13424)\rUpdating files:  77% (10337/13424)\rUpdating files:  77% (10433/13424)\rUpdating files:  78% (10471/13424)\rUpdating files:  79% (10605/13424)\rUpdating files:  80% (10740/13424)\rUpdating files:  81% (10874/13424)\rUpdating files:  82% (11008/13424)\rUpdating files:  83% (11142/13424)\rUpdating files:  84% (11277/13424)\rUpdating files:  85% (11411/13424)\rUpdating files:  85% (11439/13424)\rUpdating files:  86% (11545/13424)\rUpdating files:  87% (11679/13424)\rUpdating files:  88% (11814/13424)\rUpdating files:  89% (11948/13424)\rUpdating files:  90% (12082/13424)\rUpdating files:  91% (12216/13424)\rUpdating files:  92% (12351/13424)\rUpdating files:  93% (12485/13424)\rUpdating files:  94% (12619/13424)\rUpdating files:  95% (12753/13424)\rUpdating files:  96% (12888/13424)\rUpdating files:  97% (13022/13424)\rUpdating files:  98% (13156/13424)\rUpdating files:  99% (13290/13424)\rUpdating files:  99% (13303/13424)\rUpdating files: 100% (13424/13424)\rUpdating files: 100% (13424/13424), done.\n        cluster.go:125: Note: switching to 'lts-4081.3.7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at a4738dc8a7 New version: lts-4081.3.7\n        cluster.go:125: + emerge --getbinpkg --verbose coreos-sources\n        cluster.go:125: \n        cluster.go:125: !!! Your current profile is deprecated and not supported anymore.\n        cluster.go:125: !!! Use eselect profile to update your profile.\n        cluster.go:125: !!! Please upgrade to the following profile if possible:\n        cluster.go:125: \n        cluster.go:125:         default/linux/amd64/23.0/split-usr/no-multilib/hardened\n        cluster.go:125: \n        cluster.go:125: To upgrade do the following steps:\n        cluster.go:125: \n        cluster.go:125: A profile upgrade to version 23.0 is available for your architecture.\n        cluster.go:125: The new 23.0 profiles enable some toolchain hardening features and \n        cluster.go:125: performance enhancements by default, and standardize settings.\n        cluster.go:125: You can find the list of changes on the wiki tracking page [1].\n        cluster.go:125: \n        cluster.go:125: Upgrade instructions\n        cluster.go:125: \n        cluster.go:125: Note 1: If you have manually changed your CHOST to a value different from \n        cluster.go:125: what the stages and profiles set, you may have to do that in the future too.\n        cluster.go:125: In that case you should know what you are doing, hopefully; please read the \n        cluster.go:125: instructions with a critical eye then.\n        cluster.go:125: \n        cluster.go:125: Note 2: In case you are already familiar with binary packages, you should be\n        cluster.go:125: able to add \"--getbinpkg\" to the emerge calls to speed things up.\n        cluster.go:125: The use of binary packages is completely optional though, and also not\n        cluster.go:125: as much tested as the source-based upgrade path yet.\n        cluster.go:125: \n        cluster.go:125: 1. Ensure your system backups are up to date. Please also update\n        cluster.go:125:    your system fully and depclean before proceeding.\n        cluster.go:125:    glibc older than 2.36 and musl older than 1.2.4 is not supported anymore.\n        cluster.go:125: \n        cluster.go:125: 2. If you are still using one of the long-deprecated amd64 17.0 profiles \n        cluster.go:125:    (other than x32 or musl), then first complete the migration to the \n        cluster.go:125:    corresponding 17.1 profile. Instructions can be found at [3].\n        cluster.go:125:    \n        cluster.go:125: 3. If you are currently using systemd in a split-usr configuration, then first \n        cluster.go:125:    complete the migration to the corresponding merged-usr profile of the \n        cluster.go:125:    same profile version. Details on how to do this can be found in the news \n        cluster.go:125:    item [4].\n        cluster.go:125:    If you are currently using openrc, migrate to 23.0 first, keeping your disk\n        cluster.go:125:    layout. If you want to move from split-usr to merged-usr, do that afterwards.\n        cluster.go:125: \n        cluster.go:125: 4. Run \"emerge --info\" and note down the value of the CHOST variable.\n        cluster.go:125: \n        cluster.go:125: 5. Edit /etc/portage/make.conf; if there is a line defining the CHOST variable,\n        cluster.go:125:    remove it. Also delete all lines defining CHOST_... variables.\n        cluster.go:125: \n        cluster.go:125: 6. Select the 23.0 profile corresponding to your current profile, either using\n        cluster.go:125:    \"eselect profile\" or by manually setting the profile symlink.\n        cluster.go:125:    Note that old profiles are by default split-usr and the 23.0 profiles by\n        cluster.go:125:    default merged-usr. Do NOT change directory scheme now, since this will\n        cluster.go:125:    mess up your system! \n        cluster.go:125:    Instead, make sure that the new profile has the same property: for example, \n        cluster.go:125:    OLD default/linux/amd64/17.1  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/split-usr\n        cluster.go:125:              (added \"split-usr\")\n        cluster.go:125:    OLD default/linux/amd64/17.1/systemd/merged-usr  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/systemd\n        cluster.go:125:              (removed \"merged-usr\")\n        cluster.go:125:    A detailed table of the upgrade paths can be found at [5]. Please consult it.\n        cluster.go:125:    In some cases (hppa, x86) the table will tell you to pick between two choices. \n        cluster.go:125:    What you need should be obvious from your *old* CHOST value (from step 4).\n        cluster.go:125: \n        cluster.go:125: 7. Delete the contents of your binary package cache at ${PKGDIR}\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 8. In the file or directory /etc/portage/binrepos.conf (if existing), update\n        cluster.go:125:    the URI in all configuration such that they point to 23.0 profile binhost \n        cluster.go:125:    directories. The exact paths can be found in the table at [5], too.\n        cluster.go:125: \n        cluster.go:125: 9. Rebuild or reinstall from binary (if available) the following packages in\n        cluster.go:125:    this order, with the same version as already active:\n        cluster.go:125:      emerge --ask --oneshot sys-devel/binutils\n        cluster.go:125:    (you may have to run binutils-config and re-select your binutils now)\n        cluster.go:125:      emerge --ask --oneshot sys-devel/gcc\n        cluster.go:125:    (IMPORTANT: If this command wants to rebuild glibc first, do *not* let it do \n        cluster.go:125:     that; instead, abort and try again with --nodeps added to the command line.)\n        cluster.go:125:    (you may have to run gcc-config and re-select your gcc now)\n        cluster.go:125:    and the C library, i.e. for glibc-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/glibc\n        cluster.go:125:    or for musl-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/musl\n        cluster.go:125: \n        cluster.go:125: 10. Re-run \"emerge --info\" and check if CHOST has changed compared to step 4.\n        cluster.go:125: \n        cluster.go:125: If the CHOST has NOT changed, skip to step 13 (env-update). Otherwise, \n        cluster.go:125: \n        cluster.go:125: 11. Recheck with binutils-config and gcc-config that valid installed versions\n        cluster.go:125:    of binutils and gcc are selected.\n        cluster.go:125: \n        cluster.go:125: 12. Check /etc/env.d, /etc/env.d/binutils, and /etc/env.d/gcc for files that\n        cluster.go:125:    refer to the *OLD* CHOST value, and remove them. \n        cluster.go:125:    Examples how to do this can be found in the similar procedure at [6].\n        cluster.go:125: \n        cluster.go:125: 13. Run env-update \u0026\u0026 source /etc/profile\n        cluster.go:125: \n        cluster.go:125: 14. Re-emerge libtool:\n        cluster.go:125:    emerge --ask --oneshot libtool\n        cluster.go:125: \n        cluster.go:125: 15. Just for safety, delete the contents of your binary package cache at \n        cluster.go:125:     ${PKGDIR} again:\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 16. Rebuild world:\n        cluster.go:125:    emerge --ask --emptytree @world\n        cluster.go:125: \n        cluster.go:125: [1] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_transition\n        cluster.go:125: [2] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_timeline\n        cluster.go:125: [3] https://www.gentoo.org/support/news-items/2019-06-05-amd64-17-1-profiles-are-now-stable.html\n        cluster.go:125: [4] https://www.gentoo.org/support/news-items/2022-12-01-systemd-usrmerge.html\n        cluster.go:125: [5] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_update_table\n        cluster.go:125: [6] https://wiki.gentoo.org/wiki/Changing_the_CHOST_variable#Verifying_things_work\n        cluster.go:125: \n        cluster.go:125: \n        cluster.go:125: + zcat /proc/config.gz\n        cluster.go:125: ++ nproc\n        cluster.go:125: + exec make -C /usr/src/linux -j4 modules_prepare V=1\n"}],"result":"FAIL","platform":"qemu","version":"4081.3.7"}
