{"tests":[{"name":"cl.ignition.v1.xfsroot","result":"FAIL","duration":616160489435,"output":"        harness.go:646: Cluster failed starting machines: machine \"fe469759-c30d-458e-a4e1-678de78e2929\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.5:22: connect: no route to host\n"},{"name":"sysext.custom-oem","result":"FAIL","duration":643649457716,"output":"        sysext.go:383: creating a machine failed: machine \"7ee574af-cb3c-486d-a00f-0ee914d20afa\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 7ee574af-cb3c-486d-a00f-0ee914d20afa console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 7ee574af-cb3c-486d-a00f-0ee914d20afa console\n"},{"name":"cl.network.listeners","result":"PASS","duration":652873677415,"output":""},{"name":"cl.ignition.luks","result":"FAIL","duration":698932406758,"output":"        harness.go:646: Cluster failed starting machines: machine \"ff0b83de-3775-47da-9cfe-791c910a2d3c\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine ff0b83de-3775-47da-9cfe-791c910a2d3c console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-cryp…vice\u001b[0m - Cryptography Setup for data.\r\r) on machine ff0b83de-3775-47da-9cfe-791c910a2d3c console\n"},{"name":"cl.ignition.v2_1.swap","result":"FAIL","duration":819375093996,"output":"        harness.go:646: Cluster failed starting machines: machine \"3eef2a48-ba15-41bb-92f3-2952ada37247\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 3eef2a48-ba15-41bb-92f3-2952ada37247 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 3eef2a48-ba15-41bb-92f3-2952ada37247 console\n"},{"name":"cl.ignition.v2_1.ext4checkexisting","result":"FAIL","duration":862879525369,"output":"        harness.go:646: Cluster failed starting machines: machine \"67e49932-f3ae-4dba-b1df-42ab10014a64\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 67e49932-f3ae-4dba-b1df-42ab10014a64 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 67e49932-f3ae-4dba-b1df-42ab10014a64 console\n"},{"name":"cl.ignition.instantiated.enable-unit","result":"FAIL","duration":925015783257,"output":"        harness.go:646: Cluster failed starting machines: machine \"cebcbaad-4cb9-427a-8ae1-259c9eedb9a6\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine cebcbaad-4cb9-427a-8ae1-259c9eedb9a6 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine cebcbaad-4cb9-427a-8ae1-259c9eedb9a6 console\n"},{"name":"docker.containerd-restart","result":"FAIL","duration":504593105794,"output":"        harness.go:646: Cluster failed starting machines: machine \"3ecbb5cf-3054-4bbc-8896-e6d4875b5582\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 3ecbb5cf-3054-4bbc-8896-e6d4875b5582 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 3ecbb5cf-3054-4bbc-8896-e6d4875b5582 console\n"},{"name":"coreos.ignition.resource.remote","result":"FAIL","duration":604055496201,"output":"        harness.go:646: Cluster failed starting machines: machine \"2966c07c-922c-4396-83bd-9c7cf626fb3d\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.12:22: connect: connection refused\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 2966c07c-922c-4396-83bd-9c7cf626fb3d console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 2966c07c-922c-4396-83bd-9c7cf626fb3d console\n"},{"name":"coreos.locksmith.reboot","result":"FAIL","duration":1281042346308,"output":"        harness.go:646: Cluster failed starting machines: machine \"b6566c57-11a0-4b3b-910a-9590c7947908\" failed basic checks: some systemd units failed:\n● session-3.scope             loaded failed failed Session 3 of User core\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\n● user@500.service            loaded failed failed User Manager for UID 500\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine b6566c57-11a0-4b3b-910a-9590c7947908 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine b6566c57-11a0-4b3b-910a-9590c7947908 console\n"},{"name":"docker.network-openbsd-nc","result":"FAIL","duration":629766694310,"output":"        harness.go:646: Cluster failed starting machines: machine \"783d124d-2e24-4e63-878c-1f23e23ce5a4\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.14:22: connect: no route to host\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 783d124d-2e24-4e63-878c-1f23e23ce5a4 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 783d124d-2e24-4e63-878c-1f23e23ce5a4 console\n"},{"name":"coreos.tls.fetch-urls","result":"PASS","duration":250603883259,"output":""},{"name":"cl.ignition.oem.reuse","result":"FAIL","duration":616711231456,"output":"        harness.go:646: Cluster failed starting machines: machine \"77b67c79-8125-49d5-afe9-0fe9215d76ca\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.16:22: connect: connection refused\n        harness.go:616: Found emergency shell on machine 77b67c79-8125-49d5-afe9-0fe9215d76ca console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msysroot-boot.service\u001b[0m - /sysroot/boot.\r\r) on machine 77b67c79-8125-49d5-afe9-0fe9215d76ca console\n"},{"name":"cl.ignition.oem.indirect.new","result":"FAIL","duration":275390230746,"output":"        harness.go:646: Cluster failed starting machines: machine \"93e31e01-1c66-4105-9fc2-2f989dda5541\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 93e31e01-1c66-4105-9fc2-2f989dda5541 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 93e31e01-1c66-4105-9fc2-2f989dda5541 console\n"},{"name":"kubeadm.v1.33.8.cilium.cgroupv1.base","result":"FAIL","duration":935918490826,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"a40a91bf-18fb-439d-9deb-0f20a9c3bd7c\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine a40a91bf-18fb-439d-9deb-0f20a9c3bd7c console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine a40a91bf-18fb-439d-9deb-0f20a9c3bd7c console\n"},{"name":"cl.tpm.root-cryptenroll","result":"FAIL","duration":616175577828,"output":"        tpm.go:324: machine \"c999cbcb-aa19-48d5-aef3-073bff0ac664\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.19:22: connect: no route to host\n"},{"name":"misc.fips","result":"FAIL","duration":603093529596,"output":"        harness.go:646: Cluster failed starting machines: machine \"326fda12-1526-498d-849e-fb556fbb878f\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.21:22: connect: no route to host\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 326fda12-1526-498d-849e-fb556fbb878f console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 326fda12-1526-498d-849e-fb556fbb878f console\n"},{"name":"cl.cgroupv1/CgroupMounts","result":"PASS","duration":22921921157,"output":""},{"name":"cl.cgroupv1","result":"PASS","duration":318342681607,"output":"    --- PASS: cl.cgroupv1/CgroupMounts (22.92s)\n"},{"name":"cl.tpm.nonroot","result":"FAIL","duration":929484072993,"output":"        tpm.go:324: machine \"72028ace-e333-44bc-8396-00511e395566\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 72028ace-e333-44bc-8396-00511e395566 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 72028ace-e333-44bc-8396-00511e395566 console\n"},{"name":"systemd.sysusers.gshadow","result":"FAIL","duration":1405777738321,"output":"        harness.go:646: Cluster failed starting machines: machine \"12ad163d-9da2-4907-a9ab-48fc8226427d\" failed basic checks: some systemd units failed:\n● session-3.scope             loaded failed failed Session 3 of User core\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 12ad163d-9da2-4907-a9ab-48fc8226427d console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 12ad163d-9da2-4907-a9ab-48fc8226427d console\n"},{"name":"cl.tpm.root","result":"FAIL","duration":626634448986,"output":"        tpm.go:324: machine \"366ea2bf-845d-4891-8460-12d40dd39c5f\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.25:22: connect: no route to host\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 366ea2bf-845d-4891-8460-12d40dd39c5f console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 366ea2bf-845d-4891-8460-12d40dd39c5f console\n"},{"name":"sysext.disable-docker","result":"FAIL","duration":881727590369,"output":"        harness.go:646: Cluster failed starting machines: machine \"1f253e1f-cd26-4663-9a9c-1869420f3b80\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 1f253e1f-cd26-4663-9a9c-1869420f3b80 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 1f253e1f-cd26-4663-9a9c-1869420f3b80 console\n"},{"name":"coreos.selinux.boolean","result":"FAIL","duration":625901505325,"output":"        harness.go:646: Cluster failed starting machines: machine \"e8b27184-d334-4f5a-b4db-b384020904a8\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.26:22: connect: no route to host\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine e8b27184-d334-4f5a-b4db-b384020904a8 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine e8b27184-d334-4f5a-b4db-b384020904a8 console\n"},{"name":"cl.cloudinit.multipart-mime","result":"FAIL","duration":285653500592,"output":"        harness.go:646: Cluster failed starting machines: machine \"6d81cca8-c691-4644-8640-cedd0fe03a90\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 6d81cca8-c691-4644-8640-cedd0fe03a90 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 6d81cca8-c691-4644-8640-cedd0fe03a90 console\n"},{"name":"docker.selinux","result":"PASS","duration":306250573250,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: 4bf2067f7735: Pulling fs layer\n        cluster.go:125: 4bf2067f7735: Verifying Checksum\n        cluster.go:125: 4bf2067f7735: Download complete\n        cluster.go:125: 4bf2067f7735: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: sh: can't create /opt/hello: Permission denied\n"},{"name":"cl.install.cloudinit","result":"FAIL","duration":1443129448573,"output":"        harness.go:646: Cluster failed starting machines: machine \"7dbc2801-e3df-45e8-9f14-20eed00deaa6\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 7dbc2801-e3df-45e8-9f14-20eed00deaa6 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 7dbc2801-e3df-45e8-9f14-20eed00deaa6 console\n"},{"name":"cl.sysext.fallbackdownload","result":"FAIL","duration":632510265481,"output":"        update.go:633: creating test machine: machine \"edc2cafa-e6e1-475b-ad54-794ce1937505\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.33:22: connect: no route to host\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine edc2cafa-e6e1-475b-ad54-794ce1937505 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine edc2cafa-e6e1-475b-ad54-794ce1937505 console\n"},{"name":"docker.userns","result":"FAIL","duration":617727775252,"output":"        harness.go:646: Cluster failed starting machines: machine \"a15e4984-e01d-4ba7-b2ae-8aef76bc9c5a\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.34:22: connect: no route to host\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine a15e4984-e01d-4ba7-b2ae-8aef76bc9c5a console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine a15e4984-e01d-4ba7-b2ae-8aef76bc9c5a console\n"},{"name":"coreos.ignition.systemd.enable-service","result":"FAIL","duration":607274107546,"output":"        harness.go:646: Cluster failed starting machines: machine \"dec98502-741c-4a05-8c89-1c368e11954d\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.41:22: connect: no route to host\n"},{"name":"kubeadm.v1.33.8.calico.cgroupv1.base","result":"FAIL","duration":839480651280,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node with large disk: machine \"28d46871-4a73-4dd7-b375-668d5332b8ff\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.43:22: connect: connection refused\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 28d46871-4a73-4dd7-b375-668d5332b8ff console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 28d46871-4a73-4dd7-b375-668d5332b8ff console\n"},{"name":"cl.ignition.v2.users","result":"FAIL","duration":626829653308,"output":"        harness.go:646: Cluster failed starting machines: machine \"23ff6fd4-d38f-4ab0-bf7b-d2fd9c1efbdc\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.42:22: connect: no route to host\n"},{"name":"cl.etcd-member.etcdctlv3","result":"FAIL","duration":618815277371,"output":"        harness.go:646: Cluster failed starting machines: machine \"a11b2da7-6518-48f6-a4c9-a138645c63ca\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.45:22: connect: no route to host\n"},{"name":"cl.cloudinit.basic","result":"FAIL","duration":595342338829,"output":"        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39muser-configvirtfs.…\u001b[0mud-config from /media/configvirtfs.\r\r) on machine df4ddb38-ac0d-4061-87a7-b185804e1018 console\n"},{"name":"cl.disk.raid1.root","result":"FAIL","duration":630291243065,"output":"        raid.go:231: could not reboot machine: machine \"0740a0e0-ce55-4fac-bc21-8ab25beeb9bf\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 0740a0e0-ce55-4fac-bc21-8ab25beeb9bf console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 0740a0e0-ce55-4fac-bc21-8ab25beeb9bf console\n"},{"name":"cl.ignition.oem.regular.new","result":"FAIL","duration":2034382372905,"output":"        oem.go:199: Couldn't reboot machine: machine \"e83818f4-4b7f-4c08-bc31-84e7e8a0cef0\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.35:22: connect: no route to host\n        harness.go:616: Found emergency shell on machine e83818f4-4b7f-4c08-bc31-84e7e8a0cef0 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine e83818f4-4b7f-4c08-bc31-84e7e8a0cef0 console\n"},{"name":"linux.nfs.v4","result":"FAIL","duration":2375741277675,"output":"        nfs.go:80: NFS server booted.\n        nfs.go:85: Test file \"/tmp/tmp.QNM84oMDg9\" created on server.\n        nfs.go:122: Cluster.NewMachine: machine \"3998d618-2ed2-4a3b-84dd-1ff50944a22c\" failed basic checks: some systemd units failed:\n● session-3.scope             loaded failed failed Session 3 of User core\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\n● user@500.service            loaded failed failed User Manager for UID 500\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 3998d618-2ed2-4a3b-84dd-1ff50944a22c console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 3998d618-2ed2-4a3b-84dd-1ff50944a22c console\n"},{"name":"cl.verity/verify","result":"PASS","duration":316763543341,"output":"        cluster.go:125: Success\n"},{"name":"cl.tpm.root-cryptenroll-pcr-withupdate","result":"PASS","duration":4204402461552,"output":"        cluster.go:125: New TPM2 token enrolled as key slot 1.\n        cluster.go:125: Wiped slot 2.\n"},{"name":"cl.verity/corruption","result":"PASS","duration":199570313231,"output":""},{"name":"cl.verity","result":"PASS","duration":664032391869,"output":"    --- PASS: cl.verity/verify (316.76s)\n            cluster.go:125: Success\n    --- PASS: cl.verity/corruption (199.57s)\n"},{"name":"cl.ignition.partition_on_boot_disk","result":"PASS","duration":247302426389,"output":"        filesystem.go:487: lsblk -f:\nNAME    FSTYPE   FSVER LABEL      UUID                                 FSAVAIL FSUSE% MOUNTPOINTS\nloop2   squashfs 4.0                                                                  \nloop3   squashfs 4.0                                                                  \nsr0                                                                                   \nvda                                                                                   \n|-vda1  vfat     FAT32 EFI-SYSTEM 0C46-7047                                           \n|-vda2                                                                                \n|-vda3  btrfs                     de1edd48-4571-4695-92f0-7af6e33c4e3d                \n| `-usr btrfs                     de1edd48-4571-4695-92f0-7af6e33c4e3d                /usr\n|-vda4                                                                                \n|-vda6  btrfs          OEM        7dd1319a-da93-42af-ac3b-f04d4587a8af    123M     1% /oem\n|-vda7                                                                                \n|-vda9  ext4     1.0   ROOT       e02793bf-3e0d-4c7e-b11a-92c664da7ce3    5.9G     0% /\n`-vda10 xfs            VAR        16d7dd67-e7a1-4913-87df-40260656be8c    5.2G     3% /var\n        filesystem.go:489: findmnt:\nTARGET                       SOURCE     FSTYPE     OPTIONS\n/                            /dev/vda9  ext4       rw,relatime,seclabel\n|-/boot                      systemd-1  autofs     rw,relatime,fd=33,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=5870\n|-/media                     tmpfs      tmpfs      rw,nosuid,nodev,noexec,relatime,seclabel\n|-/tmp                       tmpfs      tmpfs      rw,nosuid,nodev,seclabel,size=1239304k,nr_inodes=1048576\n|-/dev                       devtmpfs   devtmpfs   rw,nosuid,seclabel,size=4096k,nr_inodes=304235,mode=755\n| |-/dev/hugepages           hugetlbfs  hugetlbfs  rw,nosuid,nodev,relatime,seclabel,pagesize=2M\n| |-/dev/mqueue              mqueue     mqueue     rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/dev/shm                 tmpfs      tmpfs      rw,nosuid,nodev,seclabel\n| `-/dev/pts                 devpts     devpts     rw,nosuid,noexec,relatime,seclabel,gid=5,mode=620,ptmxmode=000\n|-/sys                       sysfs      sysfs      rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/fs/selinux          selinuxfs  selinuxfs  rw,nosuid,noexec,relatime\n| |-/sys/kernel/debug        debugfs    debugfs    rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/kernel/security     securityfs securityfs rw,nosuid,nodev,noexec,relatime\n| |-/sys/fs/cgroup           cgroup2    cgroup2    rw,nosuid,nodev,noexec,relatime,seclabel,nsdelegate,memory_recursiveprot\n| |-/sys/fs/pstore           pstore     pstore     rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/fs/bpf              bpf        bpf        rw,nosuid,nodev,noexec,relatime,mode=700\n| |-/sys/kernel/tracing      tracefs    tracefs    rw,nosuid,nodev,noexec,relatime,seclabel\n| |-/sys/fs/fuse/connections fusectl    fusectl    rw,nosuid,nodev,noexec,relatime\n| `-/sys/kernel/config       configfs   configfs   rw,nosuid,nodev,noexec,relatime\n|-/proc                      proc       proc       rw,nosuid,nodev,noexec,relatime\n| `-/proc/sys/fs/binfmt_misc systemd-1  autofs     rw,relatime,fd=34,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=5874\n|-/run                       tmpfs      tmpfs      rw,nosuid,nodev,seclabel,size=495724k,nr_inodes=819200,mode=755\n| `-/run/user/500            tmpfs      tmpfs      rw,nosuid,nodev,relatime,seclabel,size=247860k,nr_inodes=61965,mode=700,uid=500,gid=500\n|-/var                       /dev/vda10 xfs        rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=32k,noquota\n|-/usr                       /dev/dm-0  btrfs      ro,relatime,seclabel,rescue=nologreplay,space_cache=v2,subvolid=5,subvol=/\n| `-/usr                     sysext     overlay    ro,nodev,relatime,seclabel,lowerdir=/run/systemd/sysext/meta/usr:/run/systemd/sysext/extensions/docker-flatcar/usr:/run/systemd/sysext/extensions/containerd-flatcar/usr:/usr\n|-/etc                       overlay    overlay    rw,noatime,seclabel,lowerdir=/sysroot/usr/share/flatcar/etc,upperdir=/sysroot/etc,workdir=/sysroot/.etc-work,uuid=on,metacopy=off\n`-/oem                       /dev/vda6  btrfs      rw,nodev,relatime,seclabel,discard=async,space_cache=v2,subvolid=5,subvol=/\n        cluster.go:152: + findmnt /var\n        cluster.go:152: + cat /var/hello\n"},{"name":"coreos.auth.verify","result":"FAIL","duration":612647005164,"output":"        harness.go:646: Cluster failed starting machines: machine \"3eeec9ea-7529-44cc-8bde-da46664467d3\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 3eeec9ea-7529-44cc-8bde-da46664467d3 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 3eeec9ea-7529-44cc-8bde-da46664467d3 console\n"},{"name":"kubeadm.v1.33.8.calico.base","result":"FAIL","duration":652241432525,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: machine \"caa09114-dab9-433e-8fdf-9541357fe5aa\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine caa09114-dab9-433e-8fdf-9541357fe5aa console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine caa09114-dab9-433e-8fdf-9541357fe5aa console\n"},{"name":"cl.omaha.ping","result":"PASS","duration":54925860741,"output":""},{"name":"cl.filesystem","result":"FAIL","duration":607203516464,"output":"        harness.go:646: Cluster failed starting machines: machine \"549f0664-6f47-4244-ae6e-39c4b01d9ef6\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.51:22: connect: connection refused\n        harness.go:616: Found emergency shell on machine 549f0664-6f47-4244-ae6e-39c4b01d9ef6 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msysroot-boot.service\u001b[0m - /sysroot/boot.\r\r) on machine 549f0664-6f47-4244-ae6e-39c4b01d9ef6 console\n"},{"name":"coreos.ignition.resource.local","result":"PASS","duration":315338626519,"output":""},{"name":"docker.base","result":"FAIL","duration":400695327138,"output":"        harness.go:646: Cluster failed starting machines: machine \"02fa777a-e4ea-457f-b483-dff5daab3c29\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 02fa777a-e4ea-457f-b483-dff5daab3c29 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 02fa777a-e4ea-457f-b483-dff5daab3c29 console\n"},{"name":"coreos.ignition.groups","result":"PASS","duration":74876879231,"output":""},{"name":"cl.ignition.symlink","result":"PASS","duration":51360780591,"output":"        cluster.go:152: + readlink /etc/localtime\n"},{"name":"linux.ntp","result":"PASS","duration":594339955894,"output":""},{"name":"cl.etcd-member.v2-backup-restore","result":"FAIL","duration":569129188810,"output":"        harness.go:646: Cluster failed starting machines: machine \"99cf898e-bcac-43b4-b23d-9332c8097f98\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 99cf898e-bcac-43b4-b23d-9332c8097f98 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 99cf898e-bcac-43b4-b23d-9332c8097f98 console\n"},{"name":"coreos.ignition.resource.s3.versioned","result":"PASS","duration":90661442779,"output":""},{"name":"bpf.ig/ig","result":"FAIL","duration":530668714368,"output":"        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n        cluster.go:125: time=\"2026-04-13T21:36:42Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Get \\\"http://%2Frun%2Fdocker.sock/v1.51/containers/json?all=1\\\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\"\n        cluster.go:125: Error: fetching gadget information: initializing and preparing operators: instantiating operator \"oci\": ensuring image: pulling image (if missing) \"trace_exec:v0.50.0\": pulling image \"trace_exec:v0.50.0\": copying to local repository: failed to perform \"FetchReference\" on source: Get \"https://ghcr.io/v2/inspektor-gadget/gadget/trace_exec/manifests/v0.50.0\": net/http: TLS handshake timeout\n        ig.go:79: ig run trace_exec did not behave as expected: Process exited with status 1\n"},{"name":"bpf.ig","result":"FAIL","duration":533037289056,"output":"    --- FAIL: bpf.ig/ig (530.67s)\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n            cluster.go:125: time=\"2026-04-13T21:36:42Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Get \\\"http://%2Frun%2Fdocker.sock/v1.51/containers/json?all=1\\\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\"\n            cluster.go:125: Error: fetching gadget information: initializing and preparing operators: instantiating operator \"oci\": ensuring image: pulling image (if missing) \"trace_exec:v0.50.0\": pulling image \"trace_exec:v0.50.0\": copying to local repository: failed to perform \"FetchReference\" on source: Get \"https://ghcr.io/v2/inspektor-gadget/gadget/trace_exec/manifests/v0.50.0\": net/http: TLS handshake timeout\n            ig.go:79: ig run trace_exec did not behave as expected: Process exited with status 1\n"},{"name":"cl.flannel.udp","result":"FAIL","duration":2760873128897,"output":"        harness.go:646: Cluster failed starting machines: machine \"40a3b503-84a4-4c2b-a21b-52a2e641b4e3\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 40a3b503-84a4-4c2b-a21b-52a2e641b4e3 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 40a3b503-84a4-4c2b-a21b-52a2e641b4e3 console\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 98551281-4166-4ec8-812b-b5f7d13defa1 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 98551281-4166-4ec8-812b-b5f7d13defa1 console\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 424dc858-8b16-4274-8062-6d7640174d2f console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 424dc858-8b16-4274-8062-6d7640174d2f console\n"},{"name":"cl.ignition.translation","result":"PASS","duration":589630640696,"output":"        cluster.go:152: + ip --json address show kola | jq -r '.[] | .addr_info | .[] | select( .family == \"inet\") | .local'\n        cluster.go:152: + cat /etc/systemd/network/00-dummy.network\n"},{"name":"kubeadm.v1.35.1.flannel.base","result":"FAIL","duration":693154853507,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node with large disk: machine \"74bef571-ff33-44dd-9258-b05d170c4941\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 74bef571-ff33-44dd-9258-b05d170c4941 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 74bef571-ff33-44dd-9258-b05d170c4941 console\n"},{"name":"cl.etcd-member.discovery","result":"FAIL","duration":4244885916343,"output":"        harness.go:646: Cluster failed starting machines: machine \"46f22ad9-8ec9-4d57-9067-71b9657034e8\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 46f22ad9-8ec9-4d57-9067-71b9657034e8 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 46f22ad9-8ec9-4d57-9067-71b9657034e8 console\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine ae5ddc91-4070-4f0d-93cb-3effa2c4a0ed console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine ae5ddc91-4070-4f0d-93cb-3effa2c4a0ed console\n"},{"name":"kubeadm.v1.33.8.flannel.cgroupv1.base","result":"FAIL","duration":928794837776,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node with large disk: machine \"71ec34a8-f3f4-4fed-91a4-84ac8597ac0b\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.70:22: connect: connection refused\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 71ec34a8-f3f4-4fed-91a4-84ac8597ac0b console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 71ec34a8-f3f4-4fed-91a4-84ac8597ac0b console\n"},{"name":"cl.tang.nonroot","result":"PASS","duration":1445138248194,"output":"        tang.go:148: Started tang on 10.0.0.1:33705\n"},{"name":"cl.network.wireguard","result":"FAIL","duration":1303770102985,"output":"        harness.go:646: Cluster failed starting machines: machine \"c7be13e0-e951-4c12-93a6-388e25b453a4\" failed basic checks: some systemd units failed:\n● session-2.scope loaded failed failed Session 2 of User core\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine c7be13e0-e951-4c12-93a6-388e25b453a4 console\n"},{"name":"cl.cloudinit.script","result":"PASS","duration":254606694186,"output":""},{"name":"cl.basic/CloudConfig","result":"PASS","duration":303452049872,"output":""},{"name":"cl.update.badverity","result":"FAIL","duration":1974565249487,"output":"        update.go:168: some systemd units failed:\n● session-1.scope  loaded failed failed Session 1 of User core\n● session-4.scope  loaded failed failed Session 4 of User core\n● user@500.service loaded failed failed User Manager for UID 500\nstatus: \njournal:-- No entries --\n"},{"name":"cl.basic/Script","result":"PASS","duration":29967807472,"output":""},{"name":"cl.basic/PortSSH","result":"PASS","duration":7221421448,"output":""},{"name":"cl.basic/Symlink","result":"PASS","duration":7945221324,"output":""},{"name":"cl.basic/SymlinkFlatcar","result":"PASS","duration":18598986817,"output":""},{"name":"cl.basic/ServicesActive","result":"PASS","duration":28043938789,"output":""},{"name":"cl.tang.root","result":"PASS","duration":407804974664,"output":"        tang.go:148: Started tang on 10.0.0.1:39547\n"},{"name":"cl.basic/ReadOnly","result":"PASS","duration":79100798780,"output":""},{"name":"cl.sysext.boot","result":"PASS","duration":680204426017,"output":"        cluster.go:125: + sudo systemctl mask --now ensure-sysext\n        cluster.go:125: Created symlink /etc/systemd/system/ensure-sysext.service → /dev/null.\n        cluster.go:125: + sudo systemctl mask --now systemd-sysext\n        cluster.go:125: Created symlink /etc/systemd/system/systemd-sysext.service → /dev/null.\n        cluster.go:125: + sudo mkdir -p /etc/flatcar/sysext /etc/flatcar/oem-sysext /oem/sysext /etc/extensions\n        cluster.go:125: + sudo tee /oem/oem-release\n        cluster.go:125: + echo ID=test\n        cluster.go:125: + echo myext\n        cluster.go:125: + sudo tee /etc/flatcar/enabled-sysext.conf\n        cluster.go:125: + sudo touch /oem/sysext/active-oem-test /etc/flatcar/oem-sysext/oem-test-4081.3.7.raw /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw /etc/flatcar/sysext/flatcar-myext-1.2.3.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /etc/extensions/oem-test.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/sysext/flatcar-myext-1.2.3.raw /etc/extensions/flatcar-myext.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4081.3.7.raw = /oem/sysext/oem-test-4081.3.7.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw = /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw ']'\n        cluster.go:125: + sudo mv /oem/sysext/oem-test-4081.3.7.raw /etc/flatcar/oem-sysext/\n        cluster.go:125: + sudo mv /etc/flatcar/oem-sysext/oem-test-1.2.3.raw /oem/sysext/\n        cluster.go:125: + sudo ln -fs /oem/sysext/oem-test-1.2.3.raw /etc/extensions/oem-test.raw\n        cluster.go:125: + sudo ln -fs /etc/flatcar/sysext/flatcar-myext-1.2.3.raw /etc/extensions/flatcar-myext.raw\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4081.3.7.raw = /oem/sysext/oem-test-4081.3.7.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw = /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw ']'\n        cluster.go:125: + '[' -e /etc/flatcar/oem-sysext/oem-test-1.2.3.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/oem-test.raw\n        cluster.go:125: + '[' /oem/sysext/oem-test-4081.3.7.raw = /oem/sysext/oem-test-4081.3.7.raw ']'\n        cluster.go:125: ++ readlink -f /etc/extensions/flatcar-myext.raw\n        cluster.go:125: + '[' /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw = /etc/flatcar/sysext/flatcar-myext-4081.3.7.raw ']'\n"},{"name":"cl.ignition.v2.ext4root","result":"FAIL","duration":269084146629,"output":"        harness.go:646: Cluster failed starting machines: machine \"10dcd0b8-5a67-4ec9-9782-84793a474bf3\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 10dcd0b8-5a67-4ec9-9782-84793a474bf3 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 10dcd0b8-5a67-4ec9-9782-84793a474bf3 console\n"},{"name":"cl.basic/RandomUUID","result":"PASS","duration":326901903480,"output":""},{"name":"cl.ignition.kargs","result":"FAIL","duration":618724016228,"output":"        cluster.go:152: + cat /proc/cmdline\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 12baf149-5714-46b1-9913-ba7db15b7852 console\n"},{"name":"cl.disk.raid1.data","result":"PASS","duration":118748762731,"output":""},{"name":"docker.btrfs-storage","result":"FAIL","duration":745656883457,"output":"        harness.go:646: Cluster failed starting machines: machine \"d33df03b-c06f-4299-a378-f93eb7e1cec4\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine d33df03b-c06f-4299-a378-f93eb7e1cec4 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine d33df03b-c06f-4299-a378-f93eb7e1cec4 console\n"},{"name":"cl.basic/DbusPerms","result":"PASS","duration":280080924367,"output":""},{"name":"cl.disk.raid0.data","result":"FAIL","duration":1632549800909,"output":"        raid.go:245: could not reboot machine: machine \"5a361907-46b7-4326-ba4c-4239b7be21e3\" failed basic checks: some systemd units failed:\n● session-2.scope  loaded failed failed Session 2 of User core\n● session-3.scope  loaded failed failed Session 3 of User core\n● user@500.service loaded failed failed User Manager for UID 500\nstatus: \njournal:-- No entries --\n"},{"name":"coreos.ignition.once","result":"PASS","duration":167220272301,"output":""},{"name":"systemd.journal.user","result":"FAIL","duration":930259690157,"output":"        harness.go:646: Cluster failed starting machines: machine \"295ab6bd-497b-4986-8a10-06e0bb40ba2e\" failed basic checks: some systemd units failed:\n● session-2.scope loaded failed failed Session 2 of User flatcar\nstatus: \njournal:-- No entries --\n"},{"name":"cl.basic/UpdateEngineKeys","result":"PASS","duration":170222804936,"output":""},{"name":"systemd.journal.remote","result":"PASS","duration":809547477690,"output":"        cluster.go:125: Running as unit: systemd-journal-remote-client.service; invocation ID: 27615d14706748b984233b2b062888e0\n"},{"name":"sysext.simple","result":"PASS","duration":146420833310,"output":""},{"name":"cl.tpm.root-cryptenroll-pcr-noupdate","result":"FAIL","duration":1312521634518,"output":"        tpm.go:324: machine \"a3964dc9-8163-4c53-90fd-589f865d079c\" failed basic checks: some systemd units failed:\n● session-2.scope             loaded failed failed Session 2 of User core\n● session-3.scope             loaded failed failed Session 3 of User core\n● ldconfig.service            loaded failed failed Rebuild Dynamic Linker Cache\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\n● user@500.service            loaded failed failed User Manager for UID 500\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine a3964dc9-8163-4c53-90fd-589f865d079c console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine a3964dc9-8163-4c53-90fd-589f865d079c console\n"},{"name":"cl.ignition.v1.once","result":"PASS","duration":183009648552,"output":""},{"name":"cl.overlay.cleanup","result":"PASS","duration":173566113657,"output":""},{"name":"cl.basic/Useradd","result":"PASS","duration":319775974847,"output":""},{"name":"cl.basic/MachineID","result":"PASS","duration":64086996532,"output":""},{"name":"cl.flannel.vxlan","result":"PASS","duration":953901577398,"output":"        flannel.go:121: ping from 4da014bb-7ff9-47e8-bd6e-92c0820339db(10.254.9.0) to 71f597bb-7e32-4cf9-90d5-0e97e185b9a6(10.254.22.0)\n"},{"name":"cl.basic/Microcode","result":"PASS","duration":44212410521,"output":""},{"name":"cl.basic","result":"PASS","duration":2721733265309,"output":"    --- PASS: cl.basic/CloudConfig (303.45s)\n    --- PASS: cl.basic/Script (29.97s)\n    --- PASS: cl.basic/PortSSH (7.22s)\n    --- PASS: cl.basic/Symlink (7.95s)\n    --- PASS: cl.basic/SymlinkFlatcar (18.60s)\n    --- PASS: cl.basic/ServicesActive (28.04s)\n    --- PASS: cl.basic/ReadOnly (79.10s)\n    --- PASS: cl.basic/RandomUUID (326.90s)\n    --- PASS: cl.basic/DbusPerms (280.08s)\n    --- PASS: cl.basic/UpdateEngineKeys (170.22s)\n    --- PASS: cl.basic/Useradd (319.78s)\n    --- PASS: cl.basic/MachineID (64.09s)\n    --- PASS: cl.basic/Microcode (44.21s)\n"},{"name":"coreos.ignition.security.tls","result":"FAIL","duration":575026403785,"output":"        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 31fbc967-ecd5-4315-ab4b-3f6b7b4b90d4 console\n"},{"name":"cl.network.iptables","result":"FAIL","duration":712306694977,"output":"        harness.go:646: Cluster failed starting machines: machine \"0dcf7de7-06aa-4d13-968a-2b5172e8caa8\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 0dcf7de7-06aa-4d13-968a-2b5172e8caa8 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 0dcf7de7-06aa-4d13-968a-2b5172e8caa8 console\n"},{"name":"cl.ignition.v1.users","result":"PASS","duration":691154199740,"output":""},{"name":"sysext.disable-containerd","result":"PASS","duration":528418533635,"output":""},{"name":"coreos.update.badusr","result":"PASS","duration":1089210661351,"output":""},{"name":"cl.ignition.v1.ext4root","result":"PASS","duration":88661853889,"output":""},{"name":"kubeadm.v1.35.1.calico.base","result":"FAIL","duration":1033214725513,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node with large disk: machine \"10390f37-cc67-4219-b1e6-0f3549bbfe02\" failed to start: ssh journalctl failed: time limit exceeded: dial tcp 10.0.0.106:22: connect: no route to host\n"},{"name":"kubeadm.v1.33.8.cilium.base","result":"FAIL","duration":1503263643211,"output":"        cluster.go:125: W0413 22:41:37.646387    1948 version.go:109] could not fetch a Kubernetes version from the internet: unable to get URL \"https://dl.k8s.io/release/stable-1.txt\": Get \"https://dl.k8s.io/release/stable-1.txt\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\n        cluster.go:125: W0413 22:41:37.856381    1948 version.go:110] falling back to the local client version: v1.33.8\n        cluster.go:125: failed to create new CRI runtime service: validate service connection: validate CRI v1 runtime API for endpoint \"unix:///var/run/containerd/containerd.sock\": rpc error: code = DeadlineExceeded desc = context deadline exceeded\n        cluster.go:125: To see the stack trace of this error execute with --v=5 or higher\n        kubeadm.go:197: unable to setup cluster: unable to run master script: Process exited with status 1\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 601a30fb-953e-4b08-918f-9a24e82e1026 console\n"},{"name":"kubeadm.v1.33.8.flannel.base","result":"FAIL","duration":1016014230987,"output":"        kubeadm.go:197: unable to setup cluster: unable to get etcd node health: health polling failed: Process exited with status 1: \n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39metcd-member.servic…\u001b[0mtcd (System Application Container).\r\r) on machine c6862b55-4b5d-4a20-b3aa-ad796ffbee3e console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine c6862b55-4b5d-4a20-b3aa-ad796ffbee3e console\n"},{"name":"coreos.ignition.sethostname","result":"FAIL","duration":558619886805,"output":"        cluster.go:125: Failed to query system properties: Connection timed out\n        cluster.go:145: \"hostnamectl\" failed: output , status Process exited with status 1\n"},{"name":"cl.ignition.oem.indirect","result":"FAIL","duration":478685107379,"output":"        harness.go:646: Cluster failed starting machines: machine \"46d1277c-7506-4f5e-804e-52ad869c3f94\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 46d1277c-7506-4f5e-804e-52ad869c3f94 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 46d1277c-7506-4f5e-804e-52ad869c3f94 console\n"},{"name":"coreos.selinux.enforce","result":"FAIL","duration":992448666167,"output":"        selinux.go:115: failed to reboot machine: machine \"97302a89-e34d-481a-87aa-ad7846090dc0\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 97302a89-e34d-481a-87aa-ad7846090dc0 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 97302a89-e34d-481a-87aa-ad7846090dc0 console\n"},{"name":"cl.ignition.oem.wipe","result":"PASS","duration":110615030046,"output":""},{"name":"cl.swap_activation","result":"FAIL","duration":1462338619574,"output":"        harness.go:646: Cluster failed starting machines: machine \"4cafee44-6e01-41d4-a5eb-f66c81e61e6e\" failed basic checks: some systemd units failed:\n● session-3.scope  loaded failed failed Session 3 of User core\n● user@500.service loaded failed failed User Manager for UID 500\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 4cafee44-6e01-41d4-a5eb-f66c81e61e6e console\n"},{"name":"cl.update.reboot","result":"PASS","duration":371640501394,"output":""},{"name":"cl.ignition.v2_1.vfat","result":"PASS","duration":45788642033,"output":""},{"name":"cl.ignition.v2.xfsroot","result":"PASS","duration":45314778642,"output":""},{"name":"kubeadm.v1.35.1.cilium.base","result":"FAIL","duration":881658417336,"output":"        cluster.go:125: W0413 22:56:27.973257    1932 version.go:108] could not fetch a Kubernetes version from the internet: unable to get URL \"https://dl.k8s.io/release/stable-1.txt\": Get \"https://dl.k8s.io/release/stable-1.txt\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\n        cluster.go:125: W0413 22:56:27.998473    1932 version.go:109] falling back to the local client version: v1.35.1\n        cluster.go:125: error: failed to create new CRI runtime service: validate service connection: validate CRI v1 runtime API for endpoint \"unix:///var/run/containerd/containerd.sock\": rpc error: code = DeadlineExceeded desc = context deadline exceeded\n        cluster.go:125: To see the stack trace of this error execute with --v=5 or higher\n        kubeadm.go:197: unable to setup cluster: unable to run master script: Process exited with status 1\n"},{"name":"cl.users.shells","result":"PASS","duration":111262349782,"output":""},{"name":"cl.network.initramfs.second-boot","result":"PASS","duration":450206761643,"output":""},{"name":"coreos.locksmith.tls","result":"PASS","duration":188468584657,"output":""},{"name":"linux.nfs.v3","result":"PASS","duration":416332027096,"output":"        nfs.go:80: NFS server booted.\n        nfs.go:85: Test file \"/tmp/tmp.G0ITiGsGFk\" created on server.\n        nfs.go:125: NFS client booted.\n        nfs.go:133: Got NFS mount.\n"},{"name":"cl.ignition.v1.groups","result":"PASS","duration":139474037221,"output":""},{"name":"cl.disk.raid0.root","result":"PASS","duration":200750210553,"output":""},{"name":"cl.internet/NTPDate","result":"PASS","duration":8986812467,"output":""},{"name":"cl.internet/UpdateEngine","result":"PASS","duration":1600111041,"output":""},{"name":"docker.enable-service.sysext","result":"PASS","duration":47889566718,"output":"        cluster.go:152: + systemctl is-enabled docker\n"},{"name":"cl.internet/DockerPing","result":"PASS","duration":24997719961,"output":""},{"name":"cl.internet/DockerEcho","result":"PASS","duration":2848791731,"output":""},{"name":"cl.internet","result":"PASS","duration":174896087994,"output":"    --- PASS: cl.internet/NTPDate (8.99s)\n    --- PASS: cl.internet/UpdateEngine (1.60s)\n    --- PASS: cl.internet/DockerPing (25.00s)\n    --- PASS: cl.internet/DockerEcho (2.85s)\n"},{"name":"docker.lib-coreos-dockerd-compat/docker-info","result":"PASS","duration":27182055509,"output":""},{"name":"cl.ignition.v2.btrfsroot","result":"PASS","duration":366230480688,"output":""},{"name":"docker.lib-coreos-dockerd-compat/resources","result":"PASS","duration":108372350168,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.3s done\n        cluster.go:125: #1 DONE 1.2s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B 0.0s done\n        cluster.go:125: #2 DONE 0.6s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.20MB 0.7s done\n        cluster.go:125: #3 DONE 1.3s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.7s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 0.8s done\n        cluster.go:125: #5 writing image sha256:55d29a34300f3b2a8c83dd36bc0d87e9e7963a947ec517cbefb336957e8b2326\n        cluster.go:125: #5 writing image sha256:55d29a34300f3b2a8c83dd36bc0d87e9e7963a947ec517cbefb336957e8b2326 0.2s done\n        cluster.go:125: #5 naming to docker.io/library/sleep 0.1s done\n        cluster.go:125: #5 DONE 1.7s\n        cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n        cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n"},{"name":"sysext.custom-docker.sysext","result":"FAIL","duration":764817967886,"output":"        cluster.go:125: bash: line 1: docker: command not found\n        cluster.go:125: Cloning into 'sysext-bakery'...\n        cluster.go:125: Updating files:   4% (8/189)\rUpdating files:   5% (10/189)\rUpdating files:   6% (12/189)\rUpdating files:   6% (13/189)\rUpdating files:   7% (14/189)\rUpdating files:   8% (16/189)\rUpdating files:   9% (18/189)\rUpdating files:  10% (19/189)\rUpdating files:  11% (21/189)\rUpdating files:  12% (23/189)\rUpdating files:  12% (24/189)\rUpdating files:  13% (25/189)\rUpdating files:  14% (27/189)\rUpdating files:  15% (29/189)\rUpdating files:  15% (30/189)\rUpdating files:  16% (31/189)\rUpdating files:  17% (33/189)\rUpdating files:  17% (34/189)\rUpdating files:  18% (35/189)\rUpdating files:  19% (36/189)\rUpdating files:  19% (37/189)\rUpdating files:  20% (38/189)\rUpdating files:  21% (40/189)\rUpdating files:  22% (42/189)\rUpdating files:  23% (44/189)\rUpdating files:  24% (46/189)\rUpdating files:  25% (48/189)\rUpdating files:  26% (50/189)\rUpdating files:  27% (52/189)\rUpdating files:  28% (53/189)\rUpdating files:  29% (55/189)\rUpdating files:  30% (57/189)\rUpdating files:  31% (59/189)\rUpdating files:  32% (61/189)\rUpdating files:  33% (63/189)\rUpdating files:  34% (65/189)\rUpdating files:  35% (67/189)\rUpdating files:  36% (69/189)\rUpdating files:  37% (70/189)\rUpdating files:  38% (72/189)\rUpdating files:  39% (74/189)\rUpdating files:  40% (76/189)\rUpdating files:  41% (78/189)\rUpdating files:  41% (79/189)\rUpdating files:  42% (80/189)\rUpdating files:  43% (82/189)\rUpdating files:  44% (84/189)\rUpdating files:  45% (86/189)\rUpdating files:  46% (87/189)\rUpdating files:  46% (88/189)\rUpdating files:  47% (89/189)\rUpdating files:  48% (91/189)\rUpdating files:  49% (93/189)\rUpdating files:  50% (95/189)\rUpdating files:  51% (97/189)\rUpdating files:  52% (99/189)\rUpdating files:  53% (101/189)\rUpdating files:  54% (103/189)\rUpdating files:  55% (104/189)\rUpdating files:  56% (106/189)\rUpdating files:  57% (108/189)\rUpdating files:  58% (110/189)\rUpdating files:  59% (112/189)\rUpdating files:  60% (114/189)\rUpdating files:  61% (116/189)\rUpdating files:  62% (118/189)\rUpdating files:  63% (120/189)\rUpdating files:  64% (121/189)\rUpdating files:  65% (123/189)\rUpdating files:  66% (125/189)\rUpdating files:  67% (127/189)\rUpdating files:  68% (129/189)\rUpdating files:  69% (131/189)\rUpdating files:  70% (133/189)\rUpdating files:  71% (135/189)\rUpdating files:  71% (136/189)\rUpdating files:  72% (137/189)\rUpdating files:  73% (138/189)\rUpdating files:  74% (140/189)\rUpdating files:  75% (142/189)\rUpdating files:  76% (144/189)\rUpdating files:  77% (146/189)\rUpdating files:  78% (148/189)\rUpdating files:  79% (150/189)\rUpdating files:  80% (152/189)\rUpdating files:  81% (154/189)\rUpdating files:  82% (155/189)\rUpdating files:  83% (157/189)\rUpdating files:  84% (159/189)\rUpdating files:  85% (161/189)\rUpdating files:  86% (163/189)\rUpdating files:  87% (165/189)\rUpdating files:  88% (167/189)\rUpdating files:  89% (169/189)\rUpdating files:  90% (171/189)\rUpdating files:  91% (172/189)\rUpdating files:  92% (174/189)\rUpdating files:  93% (176/189)\rUpdating files:  94% (178/189)\rUpdating files:  95% (180/189)\rUpdating files:  95% (181/189)\rUpdating files:  96% (182/189)\rUpdating files:  97% (184/189)\rUpdating files:  98% (186/189)\rUpdating files:  99% (188/189)\rUpdating files: 100% (189/189)\rUpdating files: 100% (189/189), done.\n        cluster.go:125: Note: switching to '9850ffd5b2353f45a9b3bf4fb84f8138a149e3e7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at 9850ffd Merge pull request #31 from flatcar/t-lo/fix-docker-23-containerd-shim\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: resize2fs 1.47.1 (20-May-2024)\n        cluster.go:125: Reload daemon failed: Transport endpoint is not connected\n        cluster.go:145: \"sudo systemctl restart systemd-sysext \u0026\u0026 sudo systemctl daemon-reload\" failed: output , status Process exited with status 1\n"},{"name":"kubeadm.v1.34.4.cilium.base","result":"FAIL","duration":594957475068,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node with large disk: machine \"38a70ea8-282f-42a1-82cf-54c77ed621f7\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 38a70ea8-282f-42a1-82cf-54c77ed621f7 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39mserial-getty…S0.service\u001b[0m - Serial Getty on ttyS0.\r\r) on machine 38a70ea8-282f-42a1-82cf-54c77ed621f7 console\n"},{"name":"cl.toolbox.dnf-install","result":"FAIL","duration":443509396186,"output":"        harness.go:646: Cluster failed starting machines: machine \"7bc0bf52-eb03-4f65-aa05-de2caa80a4d7\" failed basic checks: some systemd units failed:\n● systemd-hwdb-update.service loaded failed failed Rebuild Hardware Database\n● systemd-udev-settle.service loaded failed failed Wait for udev To Complete Device Initialization\nstatus: \njournal:-- No entries --\n        harness.go:616: Found systemd unit failed to start (\u001b[0;1;39msystemd-hwdb-updat…ervice\u001b[0m - Rebuild Hardware Database.\r\r) on machine 7bc0bf52-eb03-4f65-aa05-de2caa80a4d7 console\n        harness.go:616: Found systemd dependency unit failed to start (\u001b[0;1;39msystemd-fsck…\u001b[0mem Check on /dev/disk/by-label/OEM.\r\r) on machine 7bc0bf52-eb03-4f65-aa05-de2caa80a4d7 console\n"},{"name":"kubeadm.v1.34.4.flannel.base","result":"FAIL","duration":3283045991986,"output":"        cluster.go:125: W0413 22:25:03.664191    1816 version.go:108] could not fetch a Kubernetes version from the internet: unable to get URL \"https://dl.k8s.io/release/stable-1.txt\": Get \"https://dl.k8s.io/release/stable-1.txt\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\n        cluster.go:125: W0413 22:25:03.676580    1816 version.go:109] falling back to the local client version: v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0413 22:34:49.972282    2469 version.go:260] remote version is much newer: v1.35.3; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.6\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 22:34:58.913062    2469 checks.go:827] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.103]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 20.718339893s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.103:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1m13.535657348s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 1m47.860945134s\n        cluster.go:125: [control-plane-check] kube-apiserver is not healthy after 30m0.413537821s\n        cluster.go:125: \n        cluster.go:125: A control plane component may have crashed or exited when started by the container runtime.\n        cluster.go:125: To troubleshoot, list all containers using your preferred container runtimes CLI.\n        cluster.go:125: Here is one example how you may list all running Kubernetes containers by using crictl:\n        cluster.go:125: \t- 'crictl --runtime-endpoint unix:///var/run/containerd/containerd.sock ps -a | grep kube | grep -v pause'\n        cluster.go:125: \tOnce you have found the failing container, you can inspect its logs with:\n        cluster.go:125: \t- 'crictl --runtime-endpoint unix:///var/run/containerd/containerd.sock logs CONTAINERID'\n        cluster.go:125: \n        cluster.go:125: error: error execution phase wait-control-plane: failed while waiting for the control plane to start: kube-apiserver check failed at https://10.0.0.103:6443/livez: Get \"https://10.0.0.103:6443/livez?timeout=10s\": context deadline exceeded\n        cluster.go:125: To see the stack trace of this error execute with --v=5 or higher\n        kubeadm.go:197: unable to setup cluster: unable to run master script: Process exited with status 1\n"},{"name":"packages/sys-cluster/ipvsadm","result":"PASS","duration":15235120111,"output":""},{"name":"packages/sys-block/open-iscsi","result":"PASS","duration":445385823106,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/targetcli-fb:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/targetcli-fb\n        cluster.go:125: a2318d6c47ec: Pulling fs layer\n        cluster.go:125: 3d3086a1439f: Pulling fs layer\n        cluster.go:125: a2318d6c47ec: Verifying Checksum\n        cluster.go:125: a2318d6c47ec: Download complete\n        cluster.go:125: 3d3086a1439f: Verifying Checksum\n        cluster.go:125: 3d3086a1439f: Download complete\n        cluster.go:125: a2318d6c47ec: Pull complete\n        cluster.go:125: 3d3086a1439f: Pull complete\n        cluster.go:125: Digest: sha256:b6cd65db981974e8b74938617218dd023775b969f9a059ced21e6ce6fa4763c1\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/targetcli-fb:latest\n        cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n        cluster.go:125: Created symlink /etc/systemd/system/remote-fs.target.wants/iscsi.service → /usr/lib/systemd/system/iscsi.service.\n"},{"name":"packages","result":"PASS","duration":567511995943,"output":"    --- PASS: packages/sys-cluster/ipvsadm (15.24s)\n    --- PASS: packages/sys-block/open-iscsi (445.39s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/targetcli-fb:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/targetcli-fb\n            cluster.go:125: a2318d6c47ec: Pulling fs layer\n            cluster.go:125: 3d3086a1439f: Pulling fs layer\n            cluster.go:125: a2318d6c47ec: Verifying Checksum\n            cluster.go:125: a2318d6c47ec: Download complete\n            cluster.go:125: 3d3086a1439f: Verifying Checksum\n            cluster.go:125: 3d3086a1439f: Download complete\n            cluster.go:125: a2318d6c47ec: Pull complete\n            cluster.go:125: 3d3086a1439f: Pull complete\n            cluster.go:125: Digest: sha256:b6cd65db981974e8b74938617218dd023775b969f9a059ced21e6ce6fa4763c1\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/targetcli-fb:latest\n            cluster.go:125: mke2fs 1.47.1 (20-May-2024)\n            cluster.go:125: Created symlink /etc/systemd/system/remote-fs.target.wants/iscsi.service → /usr/lib/systemd/system/iscsi.service.\n"},{"name":"cl.osreset.ignition-rerun","result":"PASS","duration":107266564281,"output":""},{"name":"cl.ignition.v1.sethostname","result":"PASS","duration":30122494777,"output":""},{"name":"cl.ignition.oem.regular","result":"PASS","duration":758202161514,"output":""},{"name":"cl.ignition.v1.btrfsroot","result":"PASS","duration":32004270638,"output":""},{"name":"kubeadm.v1.34.4.calico.base","result":"FAIL","duration":2083435820347,"output":"        cluster.go:125: W0413 23:07:04.988512    1878 version.go:108] could not fetch a Kubernetes version from the internet: unable to get URL \"https://dl.k8s.io/release/stable-1.txt\": Get \"https://dl.k8s.io/release/stable-1.txt\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\n        cluster.go:125: W0413 23:07:04.988910    1878 version.go:109] falling back to the local client version: v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: W0413 23:14:38.911545    2427 version.go:108] could not fetch a Kubernetes version from the internet: unable to get URL \"https://dl.k8s.io/release/stable-1.txt\": Get \"https://dl.k8s.io/release/stable-1.txt\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\n        cluster.go:125: W0413 23:14:38.939514    2427 version.go:109] falling back to the local client version: v1.34.4\n        cluster.go:125: [init] Using Kubernetes version: v1.34.4\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: W0413 23:14:44.099559    2427 checks.go:1046] [preflight] WARNING: Couldn't create the interface used for talking to the container runtime: failed to create new CRI runtime service: validate service connection: validate CRI v1 runtime API for endpoint \"unix:///var/run/containerd/containerd.sock\": rpc error: code = DeadlineExceeded desc = context deadline exceeded\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: W0413 23:14:58.898621    2427 checks.go:827] detected that the sandbox image \"registry.k8s.io/pause:3.8\" of the container runtime is inconsistent with that used by kubeadm. It is recommended to use \"registry.k8s.io/pause:3.10.1\" as the CRI sandbox image.\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local localhost] and IPs [10.96.0.1 10.0.0.132]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 11.04821455s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.132:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1m28.600045421s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 1m59.699057167s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 10m0.59367517s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node localhost as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: ceo0yx.1e3fmprcmve9mixq\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.132:6443 --token ceo0yx.1e3fmprcmve9mixq \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:dce4bf075b1ac992344765a82f63a5897893f95c8e6ea9fa078b9262ee572c51 \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: error: timed out waiting for the condition\n        kubeadm.go:197: unable to setup cluster: unable to run master script: Process exited with status 1\n"},{"name":"cl.locksmith.cluster","result":"PASS","duration":995517440329,"output":""},{"name":"cl.misc.falco","result":"PASS","duration":4098529161422,"output":"        cluster.go:125: Unable to find image 'falcosecurity/falco-driver-loader:master' locally\n        cluster.go:125: master: Pulling from falcosecurity/falco-driver-loader\n        cluster.go:125: 0858e4838ecf: Pulling fs layer\n        cluster.go:125: 2b9604c8ac9b: Pulling fs layer\n        cluster.go:125: 9a34fcea6483: Pulling fs layer\n        cluster.go:125: 1f55ac5619ae: Pulling fs layer\n        cluster.go:125: 677364613154: Pulling fs layer\n        cluster.go:125: de07b784e5b1: Pulling fs layer\n        cluster.go:125: dac426886d53: Pulling fs layer\n        cluster.go:125: b670aed5aa2d: Pulling fs layer\n        cluster.go:125: 1f55ac5619ae: Waiting\n        cluster.go:125: 677364613154: Waiting\n        cluster.go:125: de07b784e5b1: Waiting\n        cluster.go:125: dac426886d53: Waiting\n        cluster.go:125: b670aed5aa2d: Waiting\n        cluster.go:125: 2b9604c8ac9b: Verifying Checksum\n        cluster.go:125: 2b9604c8ac9b: Download complete\n        cluster.go:125: 1f55ac5619ae: Download complete\n        cluster.go:125: 677364613154: Verifying Checksum\n        cluster.go:125: 677364613154: Download complete\n        cluster.go:125: 0858e4838ecf: Verifying Checksum\n        cluster.go:125: 0858e4838ecf: Download complete\n        cluster.go:125: dac426886d53: Verifying Checksum\n        cluster.go:125: dac426886d53: Download complete\n        cluster.go:125: b670aed5aa2d: Verifying Checksum\n        cluster.go:125: b670aed5aa2d: Download complete\n        cluster.go:125: 9a34fcea6483: Verifying Checksum\n        cluster.go:125: 9a34fcea6483: Download complete\n        cluster.go:125: de07b784e5b1: Verifying Checksum\n        cluster.go:125: de07b784e5b1: Download complete\n        cluster.go:125: 0858e4838ecf: Pull complete\n        cluster.go:125: 2b9604c8ac9b: Pull complete\n        cluster.go:125: 9a34fcea6483: Pull complete\n        cluster.go:125: 1f55ac5619ae: Pull complete\n        cluster.go:125: 677364613154: Pull complete\n        cluster.go:125: de07b784e5b1: Pull complete\n        cluster.go:125: dac426886d53: Pull complete\n        cluster.go:125: b670aed5aa2d: Pull complete\n        cluster.go:125: Digest: sha256:0871e09c1cab800ab96d9a7e3e25118e33f7bbc45fbc53563c7933847b189535\n        cluster.go:125: Status: Downloaded newer image for falcosecurity/falco-driver-loader:master\n"},{"name":"docker.lib-coreos-dockerd-compat/networks-reliably","result":"PASS","duration":1410705513839,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile:\n        cluster.go:125: #1 transferring dockerfile: 108B 0.7s done\n        cluster.go:125: #1 DONE 1.4s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B 0.1s done\n        cluster.go:125: #2 DONE 0.8s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 4.04MB 2.9s done\n        cluster.go:125: #3 DONE 3.8s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.9s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers\n        cluster.go:125: #5 exporting layers 2.3s done\n        cluster.go:125: #5 writing image sha256:499b1a27294c1b51c2e374137e011f1dc1e0e1a183bb573e102065a32f5b69e1\n        cluster.go:125: #5 writing image sha256:499b1a27294c1b51c2e374137e011f1dc1e0e1a183bb573e102065a32f5b69e1 0.1s done\n        cluster.go:125: #5 naming to docker.io/library/ping\n        cluster.go:125: #5 naming to docker.io/library/ping 0.2s done\n        cluster.go:125: #5 DONE 3.4s\n"},{"name":"docker.lib-coreos-dockerd-compat/user-no-caps","result":"PASS","duration":3464497619,"output":"        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n        cluster.go:125: #1 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 5.06MB 0.1s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:b23c202cc586480b11359be0be75999d51b428452d0c67497622286b148ba64a done\n        cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n        cluster.go:125: #5 DONE 0.1s\n"},{"name":"docker.lib-coreos-dockerd-compat/ownership","result":"PASS","duration":12277590037,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/nginx\n        cluster.go:125: 9d11882d4b58: Pulling fs layer\n        cluster.go:125: 6a7080264fc9: Pulling fs layer\n        cluster.go:125: f176d5d8a1c1: Pulling fs layer\n        cluster.go:125: 63ef37274310: Pulling fs layer\n        cluster.go:125: 4744f2344932: Pulling fs layer\n        cluster.go:125: 39a24e95a1e9: Pulling fs layer\n        cluster.go:125: 400f8ac367e0: Pulling fs layer\n        cluster.go:125: 63ef37274310: Waiting\n        cluster.go:125: 39a24e95a1e9: Waiting\n        cluster.go:125: 400f8ac367e0: Waiting\n        cluster.go:125: 4744f2344932: Waiting\n        cluster.go:125: f176d5d8a1c1: Verifying Checksum\n        cluster.go:125: f176d5d8a1c1: Download complete\n        cluster.go:125: 9d11882d4b58: Download complete\n        cluster.go:125: 63ef37274310: Download complete\n        cluster.go:125: 4744f2344932: Verifying Checksum\n        cluster.go:125: 4744f2344932: Download complete\n        cluster.go:125: 39a24e95a1e9: Verifying Checksum\n        cluster.go:125: 39a24e95a1e9: Download complete\n        cluster.go:125: 400f8ac367e0: Download complete\n        cluster.go:125: 6a7080264fc9: Verifying Checksum\n        cluster.go:125: 6a7080264fc9: Download complete\n        cluster.go:125: 9d11882d4b58: Pull complete\n        cluster.go:125: 6a7080264fc9: Pull complete\n        cluster.go:125: f176d5d8a1c1: Pull complete\n        cluster.go:125: 63ef37274310: Pull complete\n        cluster.go:125: 4744f2344932: Pull complete\n        cluster.go:125: 39a24e95a1e9: Pull complete\n        cluster.go:125: 400f8ac367e0: Pull complete\n        cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"docker.lib-coreos-dockerd-compat","result":"PASS","duration":1637236573799,"output":"    --- PASS: docker.lib-coreos-dockerd-compat/docker-info (27.18s)\n    --- PASS: docker.lib-coreos-dockerd-compat/resources (108.37s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B 0.3s done\n            cluster.go:125: #1 DONE 1.2s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B 0.0s done\n            cluster.go:125: #2 DONE 0.6s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 2.20MB 0.7s done\n            cluster.go:125: #3 DONE 1.3s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.7s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers\n            cluster.go:125: #5 exporting layers 0.8s done\n            cluster.go:125: #5 writing image sha256:55d29a34300f3b2a8c83dd36bc0d87e9e7963a947ec517cbefb336957e8b2326\n            cluster.go:125: #5 writing image sha256:55d29a34300f3b2a8c83dd36bc0d87e9e7963a947ec517cbefb336957e8b2326 0.2s done\n            cluster.go:125: #5 naming to docker.io/library/sleep 0.1s done\n            cluster.go:125: #5 DONE 1.7s\n            cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n            cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n    --- PASS: docker.lib-coreos-dockerd-compat/networks-reliably (1410.71s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile:\n            cluster.go:125: #1 transferring dockerfile: 108B 0.7s done\n            cluster.go:125: #1 DONE 1.4s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context:\n            cluster.go:125: #2 transferring context: 2B 0.1s done\n            cluster.go:125: #2 DONE 0.8s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 4.04MB 2.9s done\n            cluster.go:125: #3 DONE 3.8s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.9s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers\n            cluster.go:125: #5 exporting layers 2.3s done\n            cluster.go:125: #5 writing image sha256:499b1a27294c1b51c2e374137e011f1dc1e0e1a183bb573e102065a32f5b69e1\n            cluster.go:125: #5 writing image sha256:499b1a27294c1b51c2e374137e011f1dc1e0e1a183bb573e102065a32f5b69e1 0.1s done\n            cluster.go:125: #5 naming to docker.io/library/ping\n            cluster.go:125: #5 naming to docker.io/library/ping 0.2s done\n            cluster.go:125: #5 DONE 3.4s\n    --- PASS: docker.lib-coreos-dockerd-compat/user-no-caps (3.46s)\n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n            cluster.go:125: #1 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 5.06MB 0.1s done\n            cluster.go:125: #3 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.1s done\n            cluster.go:125: #5 writing image sha256:b23c202cc586480b11359be0be75999d51b428452d0c67497622286b148ba64a done\n            cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n            cluster.go:125: #5 DONE 0.1s\n    --- PASS: docker.lib-coreos-dockerd-compat/ownership (12.28s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/nginx\n            cluster.go:125: 9d11882d4b58: Pulling fs layer\n            cluster.go:125: 6a7080264fc9: Pulling fs layer\n            cluster.go:125: f176d5d8a1c1: Pulling fs layer\n            cluster.go:125: 63ef37274310: Pulling fs layer\n            cluster.go:125: 4744f2344932: Pulling fs layer\n            cluster.go:125: 39a24e95a1e9: Pulling fs layer\n            cluster.go:125: 400f8ac367e0: Pulling fs layer\n            cluster.go:125: 63ef37274310: Waiting\n            cluster.go:125: 39a24e95a1e9: Waiting\n            cluster.go:125: 400f8ac367e0: Waiting\n            cluster.go:125: 4744f2344932: Waiting\n            cluster.go:125: f176d5d8a1c1: Verifying Checksum\n            cluster.go:125: f176d5d8a1c1: Download complete\n            cluster.go:125: 9d11882d4b58: Download complete\n            cluster.go:125: 63ef37274310: Download complete\n            cluster.go:125: 4744f2344932: Verifying Checksum\n            cluster.go:125: 4744f2344932: Download complete\n            cluster.go:125: 39a24e95a1e9: Verifying Checksum\n            cluster.go:125: 39a24e95a1e9: Download complete\n            cluster.go:125: 400f8ac367e0: Download complete\n            cluster.go:125: 6a7080264fc9: Verifying Checksum\n            cluster.go:125: 6a7080264fc9: Download complete\n            cluster.go:125: 9d11882d4b58: Pull complete\n            cluster.go:125: 6a7080264fc9: Pull complete\n            cluster.go:125: f176d5d8a1c1: Pull complete\n            cluster.go:125: 63ef37274310: Pull complete\n            cluster.go:125: 4744f2344932: Pull complete\n            cluster.go:125: 39a24e95a1e9: Pull complete\n            cluster.go:125: 400f8ac367e0: Pull complete\n            cluster.go:125: Digest: sha256:37746262896e4e1a260f21898a0759befa3e3bc64a33bd95f7cd1b8400a9b03b\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"devcontainer.docker","result":"PASS","duration":8265499897935,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + ARCH=amd64\n        cluster.go:125: + VERSION=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + EXPECTED_VERSION=4081.3.7\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-22813\n        cluster.go:125: + USR_SRC_DIR=/home/core/dev-container-workdir-22813/src\n        cluster.go:125: + VAR_TMP_DIR=/home/core/dev-container-workdir-22813/tmp\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-22813/src /home/core/dev-container-workdir-22813/tmp\n        cluster.go:125: ++ parted flatcar_developer_container.bin unit b print\n        cluster.go:125: ++ grep Start --after-context=1\n        cluster.go:125: ++ tail --lines=1\n        cluster.go:125: ++ head --bytes=-2\n        cluster.go:125: ++ awk '{ print $2 }'\n        cluster.go:125: + offset=2097152\n        cluster.go:125: + mkdir root\n        cluster.go:125: + sudo mount -o loop,ro,offset=2097152 flatcar_developer_container.bin root\n        cluster.go:125: + sudo tar -C root -czf dev-container-image.tar.gz .\n        cluster.go:125: + sudo umount root\n        cluster.go:125: + rm -f flatcar_developer_container.bin\n        cluster.go:125: + docker import dev-container-image.tar.gz dev-container:42\n        cluster.go:125: + sudo rm -f dev-container-image.tar.gz\n        cluster.go:125: + restorecon /home/core/dev-container-script\n        cluster.go:125: + docker run --log-driver=journald --env PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs --env EXPECTED_VERSION=4081.3.7 --mount type=bind,source=/lib/modules,target=/lib/modules,readonly=true --mount type=bind,source=/home/core/dev-container-script,target=/home/core/dev-container-script,readonly=true --mount type=bind,source=/home/core/dev-container-workdir-22813/src,target=/usr/src --mount type=bind,source=/home/core/dev-container-workdir-22813/tmp,target=/var/tmp dev-container:42 /bin/bash /home/core/dev-container-script\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + [[ 4081.3.7 != \\4\\0\\8\\1\\.\\3\\.\\7 ]]\n        cluster.go:125: + export PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + export 'FEATURES=-ipc-sandbox -network-sandbox'\n        cluster.go:125: + FEATURES='-ipc-sandbox -network-sandbox'\n        cluster.go:125: + emerge-gitclone\n        cluster.go:125: Cloning into '/var/lib/portage/scripts'...\n        cluster.go:125: Updating files:   4% (654/13424)\rUpdating files:   5% (672/13424)\rUpdating files:   6% (806/13424)\rUpdating files:   6% (827/13424)\rUpdating files:   7% (940/13424)\rUpdating files:   8% (1074/13424)\rUpdating files:   9% (1209/13424)\rUpdating files:  10% (1343/13424)\rUpdating files:  11% (1477/13424)\rUpdating files:  11% (1532/13424)\rUpdating files:  12% (1611/13424)\rUpdating files:  13% (1746/13424)\rUpdating files:  14% (1880/13424)\rUpdating files:  14% (1971/13424)\rUpdating files:  15% (2014/13424)\rUpdating files:  16% (2148/13424)\rUpdating files:  17% (2283/13424)\rUpdating files:  18% (2417/13424)\rUpdating files:  19% (2551/13424)\rUpdating files:  19% (2590/13424)\rUpdating files:  20% (2685/13424)\rUpdating files:  21% (2820/13424)\rUpdating files:  21% (2829/13424)\rUpdating files:  22% (2954/13424)\rUpdating files:  23% (3088/13424)\rUpdating files:  24% (3222/13424)\rUpdating files:  24% (3245/13424)\rUpdating files:  25% (3356/13424)\rUpdating files:  26% (3491/13424)\rUpdating files:  27% (3625/13424)\rUpdating files:  27% (3721/13424)\rUpdating files:  28% (3759/13424)\rUpdating files:  29% (3893/13424)\rUpdating files:  30% (4028/13424)\rUpdating files:  31% (4162/13424)\rUpdating files:  31% (4295/13424)\rUpdating files:  32% (4296/13424)\rUpdating files:  33% (4430/13424)\rUpdating files:  34% (4565/13424)\rUpdating files:  35% (4699/13424)\rUpdating files:  35% (4752/13424)\rUpdating files:  36% (4833/13424)\rUpdating files:  37% (4967/13424)\rUpdating files:  37% (4995/13424)\rUpdating files:  38% (5102/13424)\rUpdating files:  39% (5236/13424)\rUpdating files:  39% (5316/13424)\rUpdating files:  40% (5370/13424)\rUpdating files:  41% (5504/13424)\rUpdating files:  42% (5639/13424)\rUpdating files:  43% (5773/13424)\rUpdating files:  43% (5780/13424)\rUpdating files:  44% (5907/13424)\rUpdating files:  45% (6041/13424)\rUpdating files:  45% (6070/13424)\rUpdating files:  46% (6176/13424)\rUpdating files:  47% (6310/13424)\rUpdating files:  47% (6321/13424)\rUpdating files:  47% (6329/13424)\rUpdating files:  47% (6338/13424)\rUpdating files:  48% (6444/13424)\rUpdating files:  48% (6507/13424)\rUpdating files:  49% (6578/13424)\rUpdating files:  50% (6712/13424)\rUpdating files:  51% (6847/13424)\rUpdating files:  52% (6981/13424)\rUpdating files:  53% (7115/13424)\rUpdating files:  54% (7249/13424)\rUpdating files:  55% (7384/13424)\rUpdating files:  56% (7518/13424)\rUpdating files:  57% (7652/13424)\rUpdating files:  58% (7786/13424)\rUpdating files:  59% (7921/13424)\rUpdating files:  60% (8055/13424)\rUpdating files:  61% (8189/13424)\rUpdating files:  62% (8323/13424)\rUpdating files:  63% (8458/13424)\rUpdating files:  64% (8592/13424)\rUpdating files:  65% (8726/13424)\rUpdating files:  66% (8860/13424)\rUpdating files:  67% (8995/13424)\rUpdating files:  67% (9008/13424)\rUpdating files:  68% (9129/13424)\rUpdating files:  69% (9263/13424)\rUpdating files:  70% (9397/13424)\rUpdating files:  71% (9532/13424)\rUpdating files:  72% (9666/13424)\rUpdating files:  73% (9800/13424)\rUpdating files:  74% (9934/13424)\rUpdating files:  75% (10068/13424)\rUpdating files:  76% (10203/13424)\rUpdating files:  76% (10216/13424)\rUpdating files:  77% (10337/13424)\rUpdating files:  78% (10471/13424)\rUpdating files:  79% (10605/13424)\rUpdating files:  79% (10691/13424)\rUpdating files:  80% (10740/13424)\rUpdating files:  81% (10874/13424)\rUpdating files:  82% (11008/13424)\rUpdating files:  82% (11019/13424)\rUpdating files:  83% (11142/13424)\rUpdating files:  83% (11227/13424)\rUpdating files:  83% (11259/13424)\rUpdating files:  84% (11277/13424)\rUpdating files:  85% (11411/13424)\rUpdating files:  86% (11545/13424)\rUpdating files:  87% (11679/13424)\rUpdating files:  88% (11814/13424)\rUpdating files:  89% (11948/13424)\rUpdating files:  89% (11968/13424)\rUpdating files:  90% (12082/13424)\rUpdating files:  91% (12216/13424)\rUpdating files:  92% (12351/13424)\rUpdating files:  93% (12485/13424)\rUpdating files:  94% (12619/13424)\rUpdating files:  94% (12717/13424)\rUpdating files:  95% (12753/13424)\rUpdating files:  96% (12888/13424)\rUpdating files:  97% (13022/13424)\rUpdating files:  98% (13156/13424)\rUpdating files:  99% (13290/13424)\rUpdating files: 100% (13424/13424)\rUpdating files: 100% (13424/13424), done.\n        cluster.go:125: Note: switching to 'lts-4081.3.7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at a4738dc8a7 New version: lts-4081.3.7\n        cluster.go:125: + emerge --getbinpkg --verbose coreos-sources\n        cluster.go:125: \n        cluster.go:125: !!! Your current profile is deprecated and not supported anymore.\n        cluster.go:125: !!! Use eselect profile to update your profile.\n        cluster.go:125: !!! Please upgrade to the following profile if possible:\n        cluster.go:125: \n        cluster.go:125:         default/linux/amd64/23.0/split-usr/no-multilib/hardened\n        cluster.go:125: \n        cluster.go:125: To upgrade do the following steps:\n        cluster.go:125: \n        cluster.go:125: A profile upgrade to version 23.0 is available for your architecture.\n        cluster.go:125: The new 23.0 profiles enable some toolchain hardening features and \n        cluster.go:125: performance enhancements by default, and standardize settings.\n        cluster.go:125: You can find the list of changes on the wiki tracking page [1].\n        cluster.go:125: \n        cluster.go:125: Upgrade instructions\n        cluster.go:125: \n        cluster.go:125: Note 1: If you have manually changed your CHOST to a value different from \n        cluster.go:125: what the stages and profiles set, you may have to do that in the future too.\n        cluster.go:125: In that case you should know what you are doing, hopefully; please read the \n        cluster.go:125: instructions with a critical eye then.\n        cluster.go:125: \n        cluster.go:125: Note 2: In case you are already familiar with binary packages, you should be\n        cluster.go:125: able to add \"--getbinpkg\" to the emerge calls to speed things up.\n        cluster.go:125: The use of binary packages is completely optional though, and also not\n        cluster.go:125: as much tested as the source-based upgrade path yet.\n        cluster.go:125: \n        cluster.go:125: 1. Ensure your system backups are up to date. Please also update\n        cluster.go:125:    your system fully and depclean before proceeding.\n        cluster.go:125:    glibc older than 2.36 and musl older than 1.2.4 is not supported anymore.\n        cluster.go:125: \n        cluster.go:125: 2. If you are still using one of the long-deprecated amd64 17.0 profiles \n        cluster.go:125:    (other than x32 or musl), then first complete the migration to the \n        cluster.go:125:    corresponding 17.1 profile. Instructions can be found at [3].\n        cluster.go:125:    \n        cluster.go:125: 3. If you are currently using systemd in a split-usr configuration, then first \n        cluster.go:125:    complete the migration to the corresponding merged-usr profile of the \n        cluster.go:125:    same profile version. Details on how to do this can be found in the news \n        cluster.go:125:    item [4].\n        cluster.go:125:    If you are currently using openrc, migrate to 23.0 first, keeping your disk\n        cluster.go:125:    layout. If you want to move from split-usr to merged-usr, do that afterwards.\n        cluster.go:125: \n        cluster.go:125: 4. Run \"emerge --info\" and note down the value of the CHOST variable.\n        cluster.go:125: \n        cluster.go:125: 5. Edit /etc/portage/make.conf; if there is a line defining the CHOST variable,\n        cluster.go:125:    remove it. Also delete all lines defining CHOST_... variables.\n        cluster.go:125: \n        cluster.go:125: 6. Select the 23.0 profile corresponding to your current profile, either using\n        cluster.go:125:    \"eselect profile\" or by manually setting the profile symlink.\n        cluster.go:125:    Note that old profiles are by default split-usr and the 23.0 profiles by\n        cluster.go:125:    default merged-usr. Do NOT change directory scheme now, since this will\n        cluster.go:125:    mess up your system! \n        cluster.go:125:    Instead, make sure that the new profile has the same property: for example, \n        cluster.go:125:    OLD default/linux/amd64/17.1  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/split-usr\n        cluster.go:125:              (added \"split-usr\")\n        cluster.go:125:    OLD default/linux/amd64/17.1/systemd/merged-usr  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/systemd\n        cluster.go:125:              (removed \"merged-usr\")\n        cluster.go:125:    A detailed table of the upgrade paths can be found at [5]. Please consult it.\n        cluster.go:125:    In some cases (hppa, x86) the table will tell you to pick between two choices. \n        cluster.go:125:    What you need should be obvious from your *old* CHOST value (from step 4).\n        cluster.go:125: \n        cluster.go:125: 7. Delete the contents of your binary package cache at ${PKGDIR}\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 8. In the file or directory /etc/portage/binrepos.conf (if existing), update\n        cluster.go:125:    the URI in all configuration such that they point to 23.0 profile binhost \n        cluster.go:125:    directories. The exact paths can be found in the table at [5], too.\n        cluster.go:125: \n        cluster.go:125: 9. Rebuild or reinstall from binary (if available) the following packages in\n        cluster.go:125:    this order, with the same version as already active:\n        cluster.go:125:      emerge --ask --oneshot sys-devel/binutils\n        cluster.go:125:    (you may have to run binutils-config and re-select your binutils now)\n        cluster.go:125:      emerge --ask --oneshot sys-devel/gcc\n        cluster.go:125:    (IMPORTANT: If this command wants to rebuild glibc first, do *not* let it do \n        cluster.go:125:     that; instead, abort and try again with --nodeps added to the command line.)\n        cluster.go:125:    (you may have to run gcc-config and re-select your gcc now)\n        cluster.go:125:    and the C library, i.e. for glibc-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/glibc\n        cluster.go:125:    or for musl-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/musl\n        cluster.go:125: \n        cluster.go:125: 10. Re-run \"emerge --info\" and check if CHOST has changed compared to step 4.\n        cluster.go:125: \n        cluster.go:125: If the CHOST has NOT changed, skip to step 13 (env-update). Otherwise, \n        cluster.go:125: \n        cluster.go:125: 11. Recheck with binutils-config and gcc-config that valid installed versions\n        cluster.go:125:    of binutils and gcc are selected.\n        cluster.go:125: \n        cluster.go:125: 12. Check /etc/env.d, /etc/env.d/binutils, and /etc/env.d/gcc for files that\n        cluster.go:125:    refer to the *OLD* CHOST value, and remove them. \n        cluster.go:125:    Examples how to do this can be found in the similar procedure at [6].\n        cluster.go:125: \n        cluster.go:125: 13. Run env-update \u0026\u0026 source /etc/profile\n        cluster.go:125: \n        cluster.go:125: 14. Re-emerge libtool:\n        cluster.go:125:    emerge --ask --oneshot libtool\n        cluster.go:125: \n        cluster.go:125: 15. Just for safety, delete the contents of your binary package cache at \n        cluster.go:125:     ${PKGDIR} again:\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 16. Rebuild world:\n        cluster.go:125:    emerge --ask --emptytree @world\n        cluster.go:125: \n        cluster.go:125: [1] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_transition\n        cluster.go:125: [2] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_timeline\n        cluster.go:125: [3] https://www.gentoo.org/support/news-items/2019-06-05-amd64-17-1-profiles-are-now-stable.html\n        cluster.go:125: [4] https://www.gentoo.org/support/news-items/2022-12-01-systemd-usrmerge.html\n        cluster.go:125: [5] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_update_table\n        cluster.go:125: [6] https://wiki.gentoo.org/wiki/Changing_the_CHOST_variable#Verifying_things_work\n        cluster.go:125: \n        cluster.go:125: \n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: Unable to unshare: EPERM (for FEATURES=\"pid-sandbox\")\n        cluster.go:125: + zcat /proc/config.gz\n        cluster.go:125: ++ nproc\n        cluster.go:125: + exec make -C /usr/src/linux -j4 modules_prepare V=1\n"},{"name":"devcontainer.systemd-nspawn","result":"PASS","duration":5363739793570,"output":"        cluster.go:125: + set -euo pipefail\n        cluster.go:125: + source /home/core/download-library.sh\n        cluster.go:125: + download_dev_container_image flatcar_developer_container.bin\n        cluster.go:125: + local output_bin=flatcar_developer_container.bin\n        cluster.go:125: + shift\n        cluster.go:125: + local arch version image_url bzip2cat\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo amd64\n        cluster.go:125: + arch=amd64\n        cluster.go:125: ++ source /usr/share/flatcar/release\n        cluster.go:125: +++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: +++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: +++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: ++ echo 4081.3.7\n        cluster.go:125: + version=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2 amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/images/@ARCH@/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/@VERSION@/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + image_url=http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + echo 'Fetching developer container from http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2'\n        cluster.go:125: + curl --fail --silent --show-error --location --retry-delay 1 --retry 60 --retry-connrefused --retry-max-time 60 --connect-timeout 20 --remote-name http://bincache.flatcar-linux.net/images/amd64/4081.3.7/flatcar_developer_container.bin.bz2\n        cluster.go:125: + bzip2cat=bzcat\n        cluster.go:125: + command -v lbzcat\n        cluster.go:125: + bzip2cat=lbzcat\n        cluster.go:125: ++ lbzcat flatcar_developer_container.bin.bz2\n        cluster.go:125: + cp --sparse=always /dev/fd/63 flatcar_developer_container.bin\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + ARCH=amd64\n        cluster.go:125: + VERSION=4081.3.7\n        cluster.go:125: ++ process_template http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs amd64 4081.3.7\n        cluster.go:125: ++ local template=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local arch=amd64\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local version=4081.3.7\n        cluster.go:125: ++ shift\n        cluster.go:125: ++ local result=http://bincache.flatcar-linux.net/boards/@ARCH@-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/@VERSION@/pkgs\n        cluster.go:125: ++ result=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: ++ echo http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + EXPECTED_VERSION=4081.3.7\n        cluster.go:125: + workdir=/home/core/dev-container-workdir-21526\n        cluster.go:125: + USR_SRC_DIR=/home/core/dev-container-workdir-21526/src\n        cluster.go:125: + VAR_TMP_DIR=/home/core/dev-container-workdir-21526/tmp\n        cluster.go:125: + mkdir -p /home/core/dev-container-workdir-21526/src /home/core/dev-container-workdir-21526/tmp\n        cluster.go:125: + sudo systemd-nspawn --console=pipe --setenv=PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs --setenv=EXPECTED_VERSION=4081.3.7 --bind-ro=/lib/modules --bind-ro=/home/core/dev-container-script --bind=/home/core/dev-container-workdir-21526/src:/usr/src --bind=/home/core/dev-container-workdir-21526/tmp:/var/tmp --image=flatcar_developer_container.bin --machine=flatcar-developer-container /bin/bash /home/core/dev-container-script\n        cluster.go:125: + source /usr/share/coreos/release\n        cluster.go:125: ++ FLATCAR_RELEASE_VERSION=4081.3.7\n        cluster.go:125: ++ FLATCAR_RELEASE_BOARD=amd64-usr\n        cluster.go:125: ++ FLATCAR_RELEASE_APPID='{e96281a6-d1af-4bde-9a0a-97b76e56dc57}'\n        cluster.go:125: + [[ 4081.3.7 != \\4\\0\\8\\1\\.\\3\\.\\7 ]]\n        cluster.go:125: + export PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + PORTAGE_BINHOST=http://bincache.flatcar-linux.net/boards/amd64-usr/4081.3.7/pkgs\n        cluster.go:125: + export 'FEATURES=-ipc-sandbox -network-sandbox'\n        cluster.go:125: + FEATURES='-ipc-sandbox -network-sandbox'\n        cluster.go:125: + emerge-gitclone\n        cluster.go:125: Cloning into '/var/lib/portage/scripts'...\n        cluster.go:125: Updating files:   1% (238/13424)\rUpdating files:   2% (269/13424)\rUpdating files:   3% (403/13424)\rUpdating files:   4% (537/13424)\rUpdating files:   4% (637/13424)\rUpdating files:   5% (672/13424)\rUpdating files:   6% (806/13424)\rUpdating files:   7% (940/13424)\rUpdating files:   7% (992/13424)\rUpdating files:   8% (1074/13424)\rUpdating files:   9% (1209/13424)\rUpdating files:  10% (1343/13424)\rUpdating files:  11% (1477/13424)\rUpdating files:  11% (1491/13424)\rUpdating files:  12% (1611/13424)\rUpdating files:  12% (1659/13424)\rUpdating files:  13% (1746/13424)\rUpdating files:  13% (1776/13424)\rUpdating files:  14% (1880/13424)\rUpdating files:  14% (1990/13424)\rUpdating files:  15% (2014/13424)\rUpdating files:  16% (2148/13424)\rUpdating files:  17% (2283/13424)\rUpdating files:  18% (2417/13424)\rUpdating files:  19% (2551/13424)\rUpdating files:  19% (2604/13424)\rUpdating files:  20% (2685/13424)\rUpdating files:  21% (2820/13424)\rUpdating files:  22% (2954/13424)\rUpdating files:  23% (3088/13424)\rUpdating files:  23% (3183/13424)\rUpdating files:  24% (3222/13424)\rUpdating files:  25% (3356/13424)\rUpdating files:  25% (3455/13424)\rUpdating files:  26% (3491/13424)\rUpdating files:  27% (3625/13424)\rUpdating files:  28% (3759/13424)\rUpdating files:  29% (3893/13424)\rUpdating files:  30% (4028/13424)\rUpdating files:  30% (4046/13424)\rUpdating files:  31% (4162/13424)\rUpdating files:  32% (4296/13424)\rUpdating files:  32% (4370/13424)\rUpdating files:  33% (4430/13424)\rUpdating files:  33% (4532/13424)\rUpdating files:  34% (4565/13424)\rUpdating files:  35% (4699/13424)\rUpdating files:  36% (4833/13424)\rUpdating files:  36% (4869/13424)\rUpdating files:  37% (4967/13424)\rUpdating files:  37% (5034/13424)\rUpdating files:  38% (5102/13424)\rUpdating files:  39% (5236/13424)\rUpdating files:  39% (5279/13424)\rUpdating files:  40% (5370/13424)\rUpdating files:  40% (5400/13424)\rUpdating files:  40% (5497/13424)\rUpdating files:  41% (5504/13424)\rUpdating files:  42% (5639/13424)\rUpdating files:  42% (5706/13424)\rUpdating files:  43% (5773/13424)\rUpdating files:  43% (5833/13424)\rUpdating files:  44% (5907/13424)\rUpdating files:  44% (5967/13424)\rUpdating files:  45% (6041/13424)\rUpdating files:  45% (6046/13424)\rUpdating files:  46% (6176/13424)\rUpdating files:  47% (6310/13424)\rUpdating files:  47% (6331/13424)\rUpdating files:  48% (6444/13424)\rUpdating files:  48% (6524/13424)\rUpdating files:  49% (6578/13424)\rUpdating files:  49% (6619/13424)\rUpdating files:  49% (6693/13424)\rUpdating files:  50% (6712/13424)\rUpdating files:  50% (6819/13424)\rUpdating files:  51% (6847/13424)\rUpdating files:  52% (6981/13424)\rUpdating files:  53% (7115/13424)\rUpdating files:  54% (7249/13424)\rUpdating files:  54% (7269/13424)\rUpdating files:  55% (7384/13424)\rUpdating files:  56% (7518/13424)\rUpdating files:  56% (7578/13424)\rUpdating files:  57% (7652/13424)\rUpdating files:  58% (7786/13424)\rUpdating files:  58% (7840/13424)\rUpdating files:  59% (7921/13424)\rUpdating files:  60% (8055/13424)\rUpdating files:  60% (8107/13424)\rUpdating files:  61% (8189/13424)\rUpdating files:  62% (8323/13424)\rUpdating files:  62% (8433/13424)\rUpdating files:  63% (8458/13424)\rUpdating files:  64% (8592/13424)\rUpdating files:  64% (8711/13424)\rUpdating files:  65% (8726/13424)\rUpdating files:  66% (8860/13424)\rUpdating files:  66% (8906/13424)\rUpdating files:  67% (8995/13424)\rUpdating files:  67% (9068/13424)\rUpdating files:  68% (9129/13424)\rUpdating files:  68% (9143/13424)\rUpdating files:  68% (9239/13424)\rUpdating files:  69% (9263/13424)\rUpdating files:  69% (9319/13424)\rUpdating files:  70% (9397/13424)\rUpdating files:  70% (9477/13424)\rUpdating files:  71% (9532/13424)\rUpdating files:  71% (9553/13424)\rUpdating files:  71% (9632/13424)\rUpdating files:  72% (9666/13424)\rUpdating files:  72% (9759/13424)\rUpdating files:  73% (9800/13424)\rUpdating files:  73% (9867/13424)\rUpdating files:  74% (9934/13424)\rUpdating files:  74% (9976/13424)\rUpdating files:  75% (10068/13424)\rUpdating files:  76% (10203/13424)\rUpdating files:  77% (10337/13424)\rUpdating files:  77% (10456/13424)\rUpdating files:  78% (10471/13424)\rUpdating files:  79% (10605/13424)\rUpdating files:  79% (10606/13424)\rUpdating files:  80% (10740/13424)\rUpdating files:  80% (10846/13424)\rUpdating files:  81% (10874/13424)\rUpdating files:  81% (10958/13424)\rUpdating files:  82% (11008/13424)\rUpdating files:  82% (11075/13424)\rUpdating files:  83% (11142/13424)\rUpdating files:  84% (11277/13424)\rUpdating files:  84% (11301/13424)\rUpdating files:  85% (11411/13424)\rUpdating files:  85% (11496/13424)\rUpdating files:  86% (11545/13424)\rUpdating files:  87% (11679/13424)\rUpdating files:  88% (11814/13424)\rUpdating files:  89% (11948/13424)\rUpdating files:  89% (12021/13424)\rUpdating files:  90% (12082/13424)\rUpdating files:  91% (12216/13424)\rUpdating files:  92% (12351/13424)\rUpdating files:  93% (12485/13424)\rUpdating files:  93% (12587/13424)\rUpdating files:  94% (12619/13424)\rUpdating files:  95% (12753/13424)\rUpdating files:  96% (12888/13424)\rUpdating files:  96% (12901/13424)\rUpdating files:  96% (12998/13424)\rUpdating files:  97% (13022/13424)\rUpdating files:  98% (13156/13424)\rUpdating files:  98% (13161/13424)\rUpdating files:  98% (13252/13424)\rUpdating files:  99% (13290/13424)\rUpdating files:  99% (13398/13424)\rUpdating files: 100% (13424/13424)\rUpdating files: 100% (13424/13424), done.\n        cluster.go:125: Note: switching to 'lts-4081.3.7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at a4738dc8a7 New version: lts-4081.3.7\n        cluster.go:125: + emerge --getbinpkg --verbose coreos-sources\n        cluster.go:125: \n        cluster.go:125: !!! Your current profile is deprecated and not supported anymore.\n        cluster.go:125: !!! Use eselect profile to update your profile.\n        cluster.go:125: !!! Please upgrade to the following profile if possible:\n        cluster.go:125: \n        cluster.go:125:         default/linux/amd64/23.0/split-usr/no-multilib/hardened\n        cluster.go:125: \n        cluster.go:125: To upgrade do the following steps:\n        cluster.go:125: \n        cluster.go:125: A profile upgrade to version 23.0 is available for your architecture.\n        cluster.go:125: The new 23.0 profiles enable some toolchain hardening features and \n        cluster.go:125: performance enhancements by default, and standardize settings.\n        cluster.go:125: You can find the list of changes on the wiki tracking page [1].\n        cluster.go:125: \n        cluster.go:125: Upgrade instructions\n        cluster.go:125: \n        cluster.go:125: Note 1: If you have manually changed your CHOST to a value different from \n        cluster.go:125: what the stages and profiles set, you may have to do that in the future too.\n        cluster.go:125: In that case you should know what you are doing, hopefully; please read the \n        cluster.go:125: instructions with a critical eye then.\n        cluster.go:125: \n        cluster.go:125: Note 2: In case you are already familiar with binary packages, you should be\n        cluster.go:125: able to add \"--getbinpkg\" to the emerge calls to speed things up.\n        cluster.go:125: The use of binary packages is completely optional though, and also not\n        cluster.go:125: as much tested as the source-based upgrade path yet.\n        cluster.go:125: \n        cluster.go:125: 1. Ensure your system backups are up to date. Please also update\n        cluster.go:125:    your system fully and depclean before proceeding.\n        cluster.go:125:    glibc older than 2.36 and musl older than 1.2.4 is not supported anymore.\n        cluster.go:125: \n        cluster.go:125: 2. If you are still using one of the long-deprecated amd64 17.0 profiles \n        cluster.go:125:    (other than x32 or musl), then first complete the migration to the \n        cluster.go:125:    corresponding 17.1 profile. Instructions can be found at [3].\n        cluster.go:125:    \n        cluster.go:125: 3. If you are currently using systemd in a split-usr configuration, then first \n        cluster.go:125:    complete the migration to the corresponding merged-usr profile of the \n        cluster.go:125:    same profile version. Details on how to do this can be found in the news \n        cluster.go:125:    item [4].\n        cluster.go:125:    If you are currently using openrc, migrate to 23.0 first, keeping your disk\n        cluster.go:125:    layout. If you want to move from split-usr to merged-usr, do that afterwards.\n        cluster.go:125: \n        cluster.go:125: 4. Run \"emerge --info\" and note down the value of the CHOST variable.\n        cluster.go:125: \n        cluster.go:125: 5. Edit /etc/portage/make.conf; if there is a line defining the CHOST variable,\n        cluster.go:125:    remove it. Also delete all lines defining CHOST_... variables.\n        cluster.go:125: \n        cluster.go:125: 6. Select the 23.0 profile corresponding to your current profile, either using\n        cluster.go:125:    \"eselect profile\" or by manually setting the profile symlink.\n        cluster.go:125:    Note that old profiles are by default split-usr and the 23.0 profiles by\n        cluster.go:125:    default merged-usr. Do NOT change directory scheme now, since this will\n        cluster.go:125:    mess up your system! \n        cluster.go:125:    Instead, make sure that the new profile has the same property: for example, \n        cluster.go:125:    OLD default/linux/amd64/17.1  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/split-usr\n        cluster.go:125:              (added \"split-usr\")\n        cluster.go:125:    OLD default/linux/amd64/17.1/systemd/merged-usr  \n        cluster.go:125:         ==\u003e  NEW default/linux/amd64/23.0/systemd\n        cluster.go:125:              (removed \"merged-usr\")\n        cluster.go:125:    A detailed table of the upgrade paths can be found at [5]. Please consult it.\n        cluster.go:125:    In some cases (hppa, x86) the table will tell you to pick between two choices. \n        cluster.go:125:    What you need should be obvious from your *old* CHOST value (from step 4).\n        cluster.go:125: \n        cluster.go:125: 7. Delete the contents of your binary package cache at ${PKGDIR}\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 8. In the file or directory /etc/portage/binrepos.conf (if existing), update\n        cluster.go:125:    the URI in all configuration such that they point to 23.0 profile binhost \n        cluster.go:125:    directories. The exact paths can be found in the table at [5], too.\n        cluster.go:125: \n        cluster.go:125: 9. Rebuild or reinstall from binary (if available) the following packages in\n        cluster.go:125:    this order, with the same version as already active:\n        cluster.go:125:      emerge --ask --oneshot sys-devel/binutils\n        cluster.go:125:    (you may have to run binutils-config and re-select your binutils now)\n        cluster.go:125:      emerge --ask --oneshot sys-devel/gcc\n        cluster.go:125:    (IMPORTANT: If this command wants to rebuild glibc first, do *not* let it do \n        cluster.go:125:     that; instead, abort and try again with --nodeps added to the command line.)\n        cluster.go:125:    (you may have to run gcc-config and re-select your gcc now)\n        cluster.go:125:    and the C library, i.e. for glibc-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/glibc\n        cluster.go:125:    or for musl-based systems\n        cluster.go:125:      emerge --ask --oneshot sys-libs/musl\n        cluster.go:125: \n        cluster.go:125: 10. Re-run \"emerge --info\" and check if CHOST has changed compared to step 4.\n        cluster.go:125: \n        cluster.go:125: If the CHOST has NOT changed, skip to step 13 (env-update). Otherwise, \n        cluster.go:125: \n        cluster.go:125: 11. Recheck with binutils-config and gcc-config that valid installed versions\n        cluster.go:125:    of binutils and gcc are selected.\n        cluster.go:125: \n        cluster.go:125: 12. Check /etc/env.d, /etc/env.d/binutils, and /etc/env.d/gcc for files that\n        cluster.go:125:    refer to the *OLD* CHOST value, and remove them. \n        cluster.go:125:    Examples how to do this can be found in the similar procedure at [6].\n        cluster.go:125: \n        cluster.go:125: 13. Run env-update \u0026\u0026 source /etc/profile\n        cluster.go:125: \n        cluster.go:125: 14. Re-emerge libtool:\n        cluster.go:125:    emerge --ask --oneshot libtool\n        cluster.go:125: \n        cluster.go:125: 15. Just for safety, delete the contents of your binary package cache at \n        cluster.go:125:     ${PKGDIR} again:\n        cluster.go:125:      rm -r /var/cache/binpkgs/*\n        cluster.go:125: \n        cluster.go:125: 16. Rebuild world:\n        cluster.go:125:    emerge --ask --emptytree @world\n        cluster.go:125: \n        cluster.go:125: [1] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_transition\n        cluster.go:125: [2] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_profile_timeline\n        cluster.go:125: [3] https://www.gentoo.org/support/news-items/2019-06-05-amd64-17-1-profiles-are-now-stable.html\n        cluster.go:125: [4] https://www.gentoo.org/support/news-items/2022-12-01-systemd-usrmerge.html\n        cluster.go:125: [5] https://wiki.gentoo.org/wiki/Project:Toolchain/23.0_update_table\n        cluster.go:125: [6] https://wiki.gentoo.org/wiki/Changing_the_CHOST_variable#Verifying_things_work\n        cluster.go:125: \n        cluster.go:125: \n        cluster.go:125: + zcat /proc/config.gz\n        cluster.go:125: ++ nproc\n        cluster.go:125: + exec make -C /usr/src/linux -j4 modules_prepare V=1\n"}],"result":"FAIL","platform":"qemu","version":"4081.3.7"}
