{"tests":[{"name":"coreos.ignition.ssh.key","result":"PASS","duration":95814623322,"output":""},{"name":"cl.network.wireguard","result":"PASS","duration":95887111414,"output":"        cluster.go:152: + ip --json address show kv0 | jq -r '.[] | .addr_info | .[] | select( .family == \"inet\") | .local'\n"},{"name":"sysext.disable-containerd","result":"PASS","duration":96100762282,"output":""},{"name":"cl.ignition.v2_1.vfat","result":"PASS","duration":96773168310,"output":""},{"name":"cl.ignition.luks","result":"PASS","duration":97340159728,"output":""},{"name":"cl.cloudinit.basic","result":"PASS","duration":97614370239,"output":""},{"name":"docker.base/docker-info","result":"PASS","duration":5980471495,"output":""},{"name":"cl.basic/Script","result":"PASS","duration":1689427576,"output":""},{"name":"cl.basic/PortSSH","result":"PASS","duration":1393485412,"output":""},{"name":"sysext.simple","result":"PASS","duration":101760788592,"output":""},{"name":"docker.btrfs-storage","result":"PASS","duration":101888974219,"output":""},{"name":"cl.basic/DbusPerms","result":"PASS","duration":1395880493,"output":""},{"name":"cl.basic/Symlink","result":"PASS","duration":1390128173,"output":""},{"name":"cl.basic/SymlinkFlatcar","result":"PASS","duration":1334877106,"output":""},{"name":"cl.basic/UpdateEngineKeys","result":"PASS","duration":1371334982,"output":""},{"name":"cl.basic/ReadOnly","result":"PASS","duration":1340099879,"output":""},{"name":"cl.basic/RandomUUID","result":"PASS","duration":1478278172,"output":""},{"name":"cl.basic/CloudConfig","result":"PASS","duration":1737688537,"output":""},{"name":"docker.base/resources","result":"PASS","duration":14132207359,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n        cluster.go:125: #1 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.12MB 0.0s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.0s done\n        cluster.go:125: #5 writing image sha256:21273ba6948175caf6191ae779511e9296c661ec3522189361a718ac40992a68 done\n        cluster.go:125: #5 naming to docker.io/library/sleep done\n        cluster.go:125: #5 DONE 0.1s\n        cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n        cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n"},{"name":"cl.basic/ServicesActive","result":"PASS","duration":1360337412,"output":""},{"name":"docker.selinux","result":"PASS","duration":115082051946,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: dbc3eadfebd7: Pulling fs layer\n        cluster.go:125: dbc3eadfebd7: Verifying Checksum\n        cluster.go:125: dbc3eadfebd7: Download complete\n        cluster.go:125: dbc3eadfebd7: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: sh: can't create /opt/hello: Permission denied\n"},{"name":"cl.basic/Useradd","result":"PASS","duration":2747927532,"output":""},{"name":"cl.basic/MachineID","result":"PASS","duration":1398399313,"output":""},{"name":"cl.basic/Microcode","result":"PASS","duration":1337612429,"output":""},{"name":"cl.basic","result":"PASS","duration":121102948982,"output":"    --- PASS: cl.basic/Script (1.69s)\n    --- PASS: cl.basic/PortSSH (1.39s)\n    --- PASS: cl.basic/DbusPerms (1.40s)\n    --- PASS: cl.basic/Symlink (1.39s)\n    --- PASS: cl.basic/SymlinkFlatcar (1.33s)\n    --- PASS: cl.basic/UpdateEngineKeys (1.37s)\n    --- PASS: cl.basic/ReadOnly (1.34s)\n    --- PASS: cl.basic/RandomUUID (1.48s)\n    --- PASS: cl.basic/CloudConfig (1.74s)\n    --- PASS: cl.basic/ServicesActive (1.36s)\n    --- PASS: cl.basic/Useradd (2.75s)\n    --- PASS: cl.basic/MachineID (1.40s)\n    --- PASS: cl.basic/Microcode (1.34s)\n"},{"name":"cl.ignition.v1.users","result":"PASS","duration":135244823649,"output":""},{"name":"docker.network-openbsd-nc","result":"PASS","duration":154519826208,"output":"        docker.go:413: creating netcat containers\n        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n        cluster.go:125: #1 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.46MB 0.0s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:c50c06ad40ccf5d514f2d78012614ca2c6701999177e4e62b569a5aa14353f86 done\n        cluster.go:125: #5 naming to docker.io/library/netcat 0.0s done\n        cluster.go:125: #5 DONE 0.1s\n        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 78B 0.0s\n        cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n        cluster.go:125: #1 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.46MB 0.0s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:35943f755351a22b2d32dce5542ce67973ade534fb6cb3848197a832d240df42 done\n        cluster.go:125: #5 naming to docker.io/library/netcat 0.0s done\n        cluster.go:125: #5 DONE 0.1s\n"},{"name":"cl.verity/verify","result":"PASS","duration":66621853439,"output":"        cluster.go:125: Success\n"},{"name":"coreos.ignition.once","result":"PASS","duration":159093006647,"output":""},{"name":"cl.osreset.ignition-rerun","result":"PASS","duration":167909479417,"output":""},{"name":"cl.etcd-member.etcdctlv3","result":"PASS","duration":96947184327,"output":"        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-21T05:31:15.207179Z\",\"caller\":\"snapshot/v3_snapshot.go:65\",\"msg\":\"created temporary db file\",\"path\":\"/tmp/tmp.OWXanopp5O/snapshot.db.part\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-21T05:31:15.208117Z\",\"logger\":\"client\",\"caller\":\"v3@v3.5.18/maintenance.go:212\",\"msg\":\"opened snapshot stream; downloading\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-21T05:31:15.208142Z\",\"caller\":\"snapshot/v3_snapshot.go:73\",\"msg\":\"fetching snapshot\",\"endpoint\":\"127.0.0.1:2379\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-21T05:31:15.213191Z\",\"logger\":\"client\",\"caller\":\"v3@v3.5.18/maintenance.go:220\",\"msg\":\"completed snapshot read; closing\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-21T05:31:15.213231Z\",\"caller\":\"snapshot/v3_snapshot.go:88\",\"msg\":\"fetched snapshot\",\"endpoint\":\"127.0.0.1:2379\",\"size\":\"20 kB\",\"took\":\"now\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-21T05:31:15.213264Z\",\"caller\":\"snapshot/v3_snapshot.go:97\",\"msg\":\"saved\",\"path\":\"/tmp/tmp.OWXanopp5O/snapshot.db\"}\n        cluster.go:125: Deprecated: Use `etcdutl snapshot status` instead.\n"},{"name":"cl.ignition.translation","result":"PASS","duration":97386888549,"output":"        cluster.go:152: + ip --json address show kola | jq -r '.[] | .addr_info | .[] | select( .family == \"inet\") | .local'\n        cluster.go:152: + cat /etc/systemd/network/00-dummy.network\n"},{"name":"docker.enable-service.sysext","result":"PASS","duration":95399853464,"output":"        cluster.go:152: + systemctl is-enabled docker\n"},{"name":"cl.update.reboot","result":"PASS","duration":205251314388,"output":""},{"name":"bpf.ig/ig","result":"PASS","duration":110759979967,"output":"        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n        cluster.go:125: time=\"2026-04-21T05:31:17Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Error response from daemon: client version 1.51 is too new. Maximum supported API version is 1.50\"\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n        cluster.go:125: + docker info\n        cluster.go:125: + docker ps\n        cluster.go:125: + docker images\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n        cluster.go:125: + dig kinvolk.io\n        cluster.go:125: + dig flatcar.org\n        cluster.go:125: + dig stable.release.flatcar-linux.net\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"bpf.ig","result":"PASS","duration":113533598317,"output":"    --- PASS: bpf.ig/ig (110.76s)\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n            cluster.go:125: time=\"2026-04-21T05:31:17Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Error response from daemon: client version 1.51 is too new. Maximum supported API version is 1.50\"\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n            cluster.go:125: + docker info\n            cluster.go:125: + docker ps\n            cluster.go:125: + docker images\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n            cluster.go:125: + dig kinvolk.io\n            cluster.go:125: + dig flatcar.org\n            cluster.go:125: + dig stable.release.flatcar-linux.net\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"docker.userns","result":"PASS","duration":108230142471,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n        cluster.go:125: #1 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 2.19MB 0.0s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:f7d43bdee7cc313bae282724a0f31686090fe18f148f918d2bb13dcf32753445 done\n        cluster.go:125: #5 naming to docker.io/library/userns-test 0.0s done\n        cluster.go:125: #5 DONE 0.1s\n"},{"name":"coreos.ignition.systemd.enable-service","result":"PASS","duration":94729523123,"output":""},{"name":"cl.ignition.v1.btrfsroot","result":"PASS","duration":127225052827,"output":""},{"name":"coreos.selinux.enforce","result":"PASS","duration":135788347365,"output":""},{"name":"cl.ignition.v1.once","result":"PASS","duration":123694983568,"output":""},{"name":"sysext.custom-docker.sysext","result":"PASS","duration":143959715312,"output":"        cluster.go:125: bash: line 1: docker: command not found\n        cluster.go:125: Cloning into 'sysext-bakery'...\n        cluster.go:125: Note: switching to '9850ffd5b2353f45a9b3bf4fb84f8138a149e3e7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at 9850ffd Merge pull request #31 from flatcar/t-lo/fix-docker-23-containerd-shim\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: dbc3eadfebd7: Pulling fs layer\n        cluster.go:125: dbc3eadfebd7: Verifying Checksum\n        cluster.go:125: dbc3eadfebd7: Download complete\n        cluster.go:125: dbc3eadfebd7: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n"},{"name":"cl.ignition.kargs","result":"PASS","duration":94000975282,"output":"        cluster.go:152: + cat /proc/cmdline\n"},{"name":"docker.base/networks-reliably","result":"PASS","duration":155945060080,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 4.23MB 0.0s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:15425e32015185f7ac205d4321373d6d1af00528a2da3a32ad036c94f732a99b\n        cluster.go:125: #5 writing image sha256:15425e32015185f7ac205d4321373d6d1af00528a2da3a32ad036c94f732a99b done\n        cluster.go:125: #5 naming to docker.io/library/ping 0.0s done\n        cluster.go:125: #5 DONE 0.1s\n"},{"name":"docker.base/user-no-caps","result":"PASS","duration":3988421935,"output":"        cluster.go:125: #0 building with \"default\" instance using docker driver\n        cluster.go:125: \n        cluster.go:125: #1 [internal] load build definition from Dockerfile\n        cluster.go:125: #1 transferring dockerfile: 108B done\n        cluster.go:125: #1 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #2 [internal] load .dockerignore\n        cluster.go:125: #2 transferring context:\n        cluster.go:125: #2 transferring context: 2B done\n        cluster.go:125: #2 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #3 [internal] load build context\n        cluster.go:125: #3 transferring context: 5.15MB 0.0s done\n        cluster.go:125: #3 DONE 0.1s\n        cluster.go:125: \n        cluster.go:125: #4 [1/1] COPY . /\n        cluster.go:125: #4 DONE 0.0s\n        cluster.go:125: \n        cluster.go:125: #5 exporting to image\n        cluster.go:125: #5 exporting layers 0.1s done\n        cluster.go:125: #5 writing image sha256:b23d35de8f502fd0e7bcdfa57fa8416648a535f740beed01754aece8c10c062d done\n        cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n        cluster.go:125: #5 DONE 0.1s\n"},{"name":"docker.base/ownership","result":"PASS","duration":8491825601,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/nginx\n        cluster.go:125: f704f61efc9c: Pulling fs layer\n        cluster.go:125: 389adaee9d96: Pulling fs layer\n        cluster.go:125: 72dd4eb7fca0: Pulling fs layer\n        cluster.go:125: 21c614a32f44: Pulling fs layer\n        cluster.go:125: 6e0e56bd84b1: Pulling fs layer\n        cluster.go:125: 4291215d83f8: Pulling fs layer\n        cluster.go:125: a06e611eb27a: Pulling fs layer\n        cluster.go:125: 21c614a32f44: Waiting\n        cluster.go:125: 6e0e56bd84b1: Waiting\n        cluster.go:125: 4291215d83f8: Waiting\n        cluster.go:125: a06e611eb27a: Waiting\n        cluster.go:125: 72dd4eb7fca0: Verifying Checksum\n        cluster.go:125: 72dd4eb7fca0: Download complete\n        cluster.go:125: 21c614a32f44: Verifying Checksum\n        cluster.go:125: 21c614a32f44: Download complete\n        cluster.go:125: f704f61efc9c: Verifying Checksum\n        cluster.go:125: f704f61efc9c: Download complete\n        cluster.go:125: 389adaee9d96: Verifying Checksum\n        cluster.go:125: 389adaee9d96: Download complete\n        cluster.go:125: 6e0e56bd84b1: Verifying Checksum\n        cluster.go:125: 6e0e56bd84b1: Download complete\n        cluster.go:125: 4291215d83f8: Verifying Checksum\n        cluster.go:125: 4291215d83f8: Download complete\n        cluster.go:125: a06e611eb27a: Verifying Checksum\n        cluster.go:125: a06e611eb27a: Download complete\n        cluster.go:125: f704f61efc9c: Pull complete\n        cluster.go:125: 389adaee9d96: Pull complete\n        cluster.go:125: 72dd4eb7fca0: Pull complete\n        cluster.go:125: 21c614a32f44: Pull complete\n        cluster.go:125: 6e0e56bd84b1: Pull complete\n        cluster.go:125: 4291215d83f8: Pull complete\n        cluster.go:125: a06e611eb27a: Pull complete\n        cluster.go:125: Digest: sha256:6f170eb5f8ea893d94fe3c05cef69270f16614a99d3d922ce8a320efc43b8591\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"docker.base","result":"PASS","duration":283294049155,"output":"    --- PASS: docker.base/docker-info (5.98s)\n    --- PASS: docker.base/resources (14.13s)\n            cluster.go:125: #0 building with \"default\" instance using docker driver\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B 0.0s done\n            cluster.go:125: #1 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 2.12MB 0.0s done\n            cluster.go:125: #3 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.0s done\n            cluster.go:125: #5 writing image sha256:21273ba6948175caf6191ae779511e9296c661ec3522189361a718ac40992a68 done\n            cluster.go:125: #5 naming to docker.io/library/sleep done\n            cluster.go:125: #5 DONE 0.1s\n            cluster.go:125: WARNING: Your kernel does not support memory swappiness capabilities or the cgroup is not mounted. Memory swappiness discarded.\n            cluster.go:125: WARNING: Your kernel does not support OomKillDisable. OomKillDisable discarded.\n    --- PASS: docker.base/networks-reliably (155.95s)\n            cluster.go:125: #0 building with \"default\" instance using docker driver\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 4.23MB 0.0s done\n            cluster.go:125: #3 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.1s done\n            cluster.go:125: #5 writing image sha256:15425e32015185f7ac205d4321373d6d1af00528a2da3a32ad036c94f732a99b\n            cluster.go:125: #5 writing image sha256:15425e32015185f7ac205d4321373d6d1af00528a2da3a32ad036c94f732a99b done\n            cluster.go:125: #5 naming to docker.io/library/ping 0.0s done\n            cluster.go:125: #5 DONE 0.1s\n    --- PASS: docker.base/user-no-caps (3.99s)\n            cluster.go:125: #0 building with \"default\" instance using docker driver\n            cluster.go:125: \n            cluster.go:125: #1 [internal] load build definition from Dockerfile\n            cluster.go:125: #1 transferring dockerfile: 108B done\n            cluster.go:125: #1 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #2 [internal] load .dockerignore\n            cluster.go:125: #2 transferring context:\n            cluster.go:125: #2 transferring context: 2B done\n            cluster.go:125: #2 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #3 [internal] load build context\n            cluster.go:125: #3 transferring context: 5.15MB 0.0s done\n            cluster.go:125: #3 DONE 0.1s\n            cluster.go:125: \n            cluster.go:125: #4 [1/1] COPY . /\n            cluster.go:125: #4 DONE 0.0s\n            cluster.go:125: \n            cluster.go:125: #5 exporting to image\n            cluster.go:125: #5 exporting layers 0.1s done\n            cluster.go:125: #5 writing image sha256:b23d35de8f502fd0e7bcdfa57fa8416648a535f740beed01754aece8c10c062d done\n            cluster.go:125: #5 naming to docker.io/library/captest 0.0s done\n            cluster.go:125: #5 DONE 0.1s\n    --- PASS: docker.base/ownership (8.49s)\n            cluster.go:125: Unable to find image 'ghcr.io/flatcar/nginx:latest' locally\n            cluster.go:125: latest: Pulling from flatcar/nginx\n            cluster.go:125: f704f61efc9c: Pulling fs layer\n            cluster.go:125: 389adaee9d96: Pulling fs layer\n            cluster.go:125: 72dd4eb7fca0: Pulling fs layer\n            cluster.go:125: 21c614a32f44: Pulling fs layer\n            cluster.go:125: 6e0e56bd84b1: Pulling fs layer\n            cluster.go:125: 4291215d83f8: Pulling fs layer\n            cluster.go:125: a06e611eb27a: Pulling fs layer\n            cluster.go:125: 21c614a32f44: Waiting\n            cluster.go:125: 6e0e56bd84b1: Waiting\n            cluster.go:125: 4291215d83f8: Waiting\n            cluster.go:125: a06e611eb27a: Waiting\n            cluster.go:125: 72dd4eb7fca0: Verifying Checksum\n            cluster.go:125: 72dd4eb7fca0: Download complete\n            cluster.go:125: 21c614a32f44: Verifying Checksum\n            cluster.go:125: 21c614a32f44: Download complete\n            cluster.go:125: f704f61efc9c: Verifying Checksum\n            cluster.go:125: f704f61efc9c: Download complete\n            cluster.go:125: 389adaee9d96: Verifying Checksum\n            cluster.go:125: 389adaee9d96: Download complete\n            cluster.go:125: 6e0e56bd84b1: Verifying Checksum\n            cluster.go:125: 6e0e56bd84b1: Download complete\n            cluster.go:125: 4291215d83f8: Verifying Checksum\n            cluster.go:125: 4291215d83f8: Download complete\n            cluster.go:125: a06e611eb27a: Verifying Checksum\n            cluster.go:125: a06e611eb27a: Download complete\n            cluster.go:125: f704f61efc9c: Pull complete\n            cluster.go:125: 389adaee9d96: Pull complete\n            cluster.go:125: 72dd4eb7fca0: Pull complete\n            cluster.go:125: 21c614a32f44: Pull complete\n            cluster.go:125: 6e0e56bd84b1: Pull complete\n            cluster.go:125: 4291215d83f8: Pull complete\n            cluster.go:125: a06e611eb27a: Pull complete\n            cluster.go:125: Digest: sha256:6f170eb5f8ea893d94fe3c05cef69270f16614a99d3d922ce8a320efc43b8591\n            cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/nginx:latest\n"},{"name":"cl.ignition.v2.btrfsroot","result":"FAIL","duration":216234496,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-d437ca1297\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.update.badverity","result":"FAIL","duration":219350292,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-dca478e86c\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.symlink","result":"PASS","duration":125365127215,"output":"        cluster.go:152: + readlink /etc/localtime\n"},{"name":"cl.etcd-member.v2-backup-restore","result":"FAIL","duration":390603900,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-2019ea8b93\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.toolbox.dnf-install","result":"FAIL","duration":218616112,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-fa6040c60a\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"docker.lib-coreos-dockerd-compat","result":"FAIL","duration":284697243,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-0c4bfa2727\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.locksmith.cluster","result":"FAIL","duration":1795555862,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-0cd9fcac08\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.update.badusr","result":"FAIL","duration":280408549,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-88b59ea61f\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.35.1.calico.base","result":"FAIL","duration":2214438707,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-a681b0b791\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2_1.swap","result":"FAIL","duration":247755968,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-e569794885\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.33.8.calico.base","result":"FAIL","duration":2218484395,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-db816dffe9\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2.xfsroot","result":"FAIL","duration":258513957,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-94e12f11e6\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.network.nftables","result":"FAIL","duration":256084376,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-ad51f1b6b1\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.flannel.vxlan","result":"FAIL","duration":366570840,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-0803e1e6de\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.selinux.boolean","result":"FAIL","duration":223532825,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-62fb0364c1\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"systemd.sysusers.gshadow","result":"FAIL","duration":211162528,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-023446ce93\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.swap_activation","result":"FAIL","duration":220581078,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-03c26f82a1\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2.users","result":"FAIL","duration":222117106,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-31ca542b92\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"docker.containerd-restart","result":"FAIL","duration":208618638,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-2f21ecbf8a\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.auth.verify","result":"FAIL","duration":211315845,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-9520039d77\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.metadata.azure","result":"FAIL","duration":237589184,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-a423f3681d\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.etcd-member.discovery","result":"FAIL","duration":373720312,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-20729bddab\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.misc.empty","result":"FAIL","duration":594787770,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-1e6f71e2a3\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.noop","result":"FAIL","duration":263689033,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-4ee7a8f206\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2.noop","result":"FAIL","duration":240042691,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-f8109eae76\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.groups","result":"FAIL","duration":223549034,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-b9e6834b3d\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.users.shells","result":"FAIL","duration":702132002,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-dca3a77db6\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2.ext4root","result":"FAIL","duration":203639805,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-518271ad03\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"sysext.disable-docker","result":"FAIL","duration":211424844,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-9eef398f3f\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.cloudinit.script","result":"FAIL","duration":211445109,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-93423025b9\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.locksmith.tls","result":"FAIL","duration":264679812,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-88bbf5abdb\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.cloudinit.multipart-mime","result":"FAIL","duration":220818672,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-76c8571d11\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.overlay.cleanup","result":"FAIL","duration":86066781891,"output":"        harness.go:646: Cluster failed starting machines: machine \"ci-4676.0.0-n-82b18be005\" failed basic checks: no /etc/os-release file: ssh: handshake failed: read tcp 10.200.1.7:35558-\u003e20.83.244.18:22: read: connection reset by peer: \n"},{"name":"coreos.ignition.security.tls","result":"FAIL","duration":557951300,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-7d97c08a92\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2_1.ext4checkexisting","result":"FAIL","duration":226950635,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-437b2a67db\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.internet","result":"FAIL","duration":98541059173,"output":"        harness.go:752: dropping kolet binary: failed executing install: \"\": wait: remote command exited without exit status or exit signal\n"},{"name":"kubeadm.v1.33.8.cilium.base","result":"FAIL","duration":2240180427,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-c7b256ccc8\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.33.8.flannel.base","result":"FAIL","duration":2257606404,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Network/publicIPAddresses/ip-15643788bf\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-cd4b2527c4' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.34.4.calico.base","result":"FAIL","duration":196728596993,"output":"        kubeadm.go:197: unable to setup cluster: unable to run master script: wait: remote command exited without exit status or exit signal\n"},{"name":"systemd.journal.user","result":"FAIL","duration":92268885929,"output":"        harness.go:646: Cluster failed starting machines: waiting for machine to become active: GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Compute/virtualMachines/ci-4676.0.0-n-ea9fa82159\n--------------------------------------------------------------------------------\nRESPONSE 404: 404 Not Found\nERROR CODE: NotFound\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"NotFound\",\n    \"message\": \"The entity was not found in this Azure location.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.ignition.resource.local","result":"FAIL","duration":154058673551,"output":"        resource.go:333: starting client: PollUntilDone(ci-4676.0.0-n-60ded33f4f): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/b148f98c-d863-46f6-a23b-be7faeecc183\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-21T05:32:22.5866343+00:00\",\n  \"endTime\": \"2026-04-21T05:33:07.6606684+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"b148f98c-d863-46f6-a23b-be7faeecc183\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.ext4root","result":"FAIL","duration":103410804410,"output":"        harness.go:646: Cluster failed starting machines: waiting for machine to become active: GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Compute/virtualMachines/ci-4676.0.0-n-d3a10ed058\n--------------------------------------------------------------------------------\nRESPONSE 404: 404 Not Found\nERROR CODE: NotFound\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"NotFound\",\n    \"message\": \"The entity was not found in this Azure location.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.xfsroot","result":"FAIL","duration":102934725872,"output":"        harness.go:646: Cluster failed starting machines: waiting for machine to become active: GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Compute/virtualMachines/ci-4676.0.0-n-d632fbbdee\n--------------------------------------------------------------------------------\nRESPONSE 404: 404 Not Found\nERROR CODE: NotFound\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"NotFound\",\n    \"message\": \"The entity was not found in this Azure location.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.tls.fetch-urls","result":"FAIL","duration":93865758946,"output":"        harness.go:646: Cluster failed starting machines: waiting for machine to become active: GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-cd4b2527c4/providers/Microsoft.Compute/virtualMachines/ci-4676.0.0-n-2e5ec2260f\n--------------------------------------------------------------------------------\nRESPONSE 404: 404 Not Found\nERROR CODE: NotFound\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"NotFound\",\n    \"message\": \"The entity was not found in this Azure location.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.verity/corruption","result":"PASS","duration":175204206314,"output":""},{"name":"kubeadm.v1.34.4.flannel.base","result":"FAIL","duration":333171943612,"output":"        cluster.go:125: I0421 05:31:20.731472    2753 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0421 05:31:31.850278    3082 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4676.0.0-n-2a7d9e129b\" could not be reached\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4676.0.0-n-2a7d9e129b\": lookup ci-4676.0.0-n-2a7d9e129b on 168.63.129.16:53: no such host\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [ci-4676.0.0-n-2a7d9e129b kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 10.0.0.25]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.001499593s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.25:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 2.504175545s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 3.004473518s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 5.002397066s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node ci-4676.0.0-n-2a7d9e129b as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node ci-4676.0.0-n-2a7d9e129b as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: fmajww.yout017p0slh7cwt\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.25:6443 --token fmajww.yout017p0slh7cwt \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:0a4a9fe89603c305c442906eb04bf02ad0628447646a742fd36ec4c0d182baf3 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        kubeadm.go:197: unable to setup cluster: unable to create worker node: PollUntilDone(ci-4676.0.0-n-2bb470e65e): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/aa58cfdb-f443-4a3d-89d3-d7ce93d79069\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-21T05:31:47.3270024+00:00\",\n  \"endTime\": \"2026-04-21T05:33:22.3633336+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"aa58cfdb-f443-4a3d-89d3-d7ce93d79069\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.verity","result":"PASS","duration":335097229643,"output":"    --- PASS: cl.verity/verify (66.62s)\n            cluster.go:125: Success\n    --- PASS: cl.verity/corruption (175.20s)\n"},{"name":"kubeadm.v1.34.4.cilium.base","result":"FAIL","duration":335674711781,"output":"        cluster.go:125: I0421 05:31:26.229588    2837 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0421 05:31:36.189466    3113 version.go:260] remote version is much newer: v1.35.4; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4676.0.0-n-a44608976d\" could not be reached\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4676.0.0-n-a44608976d\": lookup ci-4676.0.0-n-a44608976d on 168.63.129.16:53: no such host\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [ci-4676.0.0-n-a44608976d kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 10.0.0.26]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.001948155s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.26:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.004074082s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 1.506167876s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 3.00163611s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node ci-4676.0.0-n-a44608976d as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node ci-4676.0.0-n-a44608976d as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: wtxywu.ucbxi8lb6qyin1qn\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.26:6443 --token wtxywu.ucbxi8lb6qyin1qn \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:952f92a3e11f7f7c71211cc54f74f692dc41bed0d5e78d4f9efe9405e21f13b5 \n        cluster.go:125: i  Using Cilium version 1.12.5\n        cluster.go:125: ? Auto-detected cluster name: kubernetes\n        cluster.go:125: ? Auto-detected datapath mode: tunnel\n        cluster.go:125: ? Auto-detected kube-proxy has been installed\n        cluster.go:125: i  helm template --namespace kube-system cilium cilium/cilium --version 1.12.5 --set cluster.id=0,cluster.name=kubernetes,encryption.nodeEncryption=false,extraConfig.cluster-pool-ipv4-cidr=192.168.0.0/17,extraConfig.enable-endpoint-routes=true,kubeProxyReplacement=disabled,operator.replicas=1,serviceAccounts.cilium.name=cilium,serviceAccounts.operator.name=cilium-operator,tunnel=vxlan\n        cluster.go:125: i  Storing helm values file in kube-system/cilium-cli-helm-values Secret\n        cluster.go:125: ? Created CA in secret cilium-ca\n        cluster.go:125: ? Generating certificates for Hubble...\n        cluster.go:125: ? Creating Service accounts...\n        cluster.go:125: ? Creating Cluster roles...\n        cluster.go:125: ? Creating ConfigMap for Cilium version 1.12.5...\n        cluster.go:125: i  Manual overwrite in ConfigMap: enable-endpoint-routes=true\n        cluster.go:125: i  Manual overwrite in ConfigMap: cluster-pool-ipv4-cidr=192.168.0.0/17\n        cluster.go:125: ? Creating Agent DaemonSet...\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/mount-cgroup]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/apply-sysctl-overwrites]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/clean-cilium-state]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: level=warning msg=\"spec.template.metadata.annotations[container.apparmor.security.beta.kubernetes.io/cilium-agent]: deprecated since v1.30; use the \\\"appArmorProfile\\\" field instead\" subsys=klog\n        cluster.go:125: ? Creating Operator Deployment...\n        cluster.go:125: ? Waiting for Cilium to be installed and ready...\n        cluster.go:125: ? Cilium was successfully installed! Run 'cilium status' to view installation health\n        cluster.go:125: \u001b[33m    /??\\\n        cluster.go:125: \u001b[36m /??\u001b[33m\\__/\u001b[32m??\\\u001b[0m    Cilium:         \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[36m \\__\u001b[31m/??\\\u001b[32m__/\u001b[0m    Operator:       \u001b[32mOK\u001b[0m\n        cluster.go:125: \u001b[32m /??\u001b[31m\\__/\u001b[35m??\\\u001b[0m    Hubble:         \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[32m \\__\u001b[34m/??\\\u001b[35m__/\u001b[0m    ClusterMesh:    \u001b[36mdisabled\u001b[0m\n        cluster.go:125: \u001b[34m    \\__/\n        cluster.go:125: \u001b[0m\n        cluster.go:125: DaemonSet        cilium             \n        cluster.go:125: Deployment       cilium-operator    \n        cluster.go:125: Containers:      cilium-operator    \n        cluster.go:125:                  cilium             \n        cluster.go:125: Cluster Pods:    0/0 managed by Cilium\n        kubeadm.go:197: unable to setup cluster: unable to create worker node: PollUntilDone(ci-4676.0.0-n-8db3db76a5): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/e047abc4-d447-4f96-b919-92c3e041ddd7\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-21T05:31:49.1687951+00:00\",\n  \"endTime\": \"2026-04-21T05:33:07.7703127+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"e047abc4-d447-4f96-b919-92c3e041ddd7\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.ignition.groups","result":"FAIL","duration":141706645202,"output":"        harness.go:646: Cluster failed starting machines: PollUntilDone(ci-4676.0.0-n-1b0b8910ea): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/6d686696-8c84-49c0-a644-8a3a829bec5c\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-21T05:31:26.2898797+00:00\",\n  \"endTime\": \"2026-04-21T05:33:13.9499716+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"6d686696-8c84-49c0-a644-8a3a829bec5c\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.ignition.resource.remote","result":"FAIL","duration":111750134894,"output":"        harness.go:646: Cluster failed starting machines: PollUntilDone(ci-4676.0.0-n-bab289ac3d): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/39d8fae6-6d9a-428f-abc7-bbefdd917753\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-21T05:32:02.0210417+00:00\",\n  \"endTime\": \"2026-04-21T05:33:22.1857034+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"39d8fae6-6d9a-428f-abc7-bbefdd917753\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.network.initramfs.second-boot","result":"FAIL","duration":771134477009,"output":"        cluster.go:145: \"journalctl -b 0 -o cat -u initrd-switch-root.target -u systemd-networkd.service\" failed: output , status dial tcp 20.69.104.166:22: i/o timeout\n"},{"name":"kubeadm.v1.35.1.flannel.base","result":"FAIL","duration":756621173474,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: PollUntilDone(ci-4676.0.0-n-a50227e771): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/f8f877ac-8ce5-4730-b506-a0371d24c834\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-21T05:32:31.3704691+00:00\",\n  \"endTime\": \"2026-04-21T05:45:00.546877+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"f8f877ac-8ce5-4730-b506-a0371d24c834\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.network.iptables","result":"FAIL","duration":784437721301,"output":"        harness.go:646: Cluster failed starting machines: PollUntilDone(ci-4676.0.0-n-bcf7131833): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/3c10c973-53e0-49f2-94be-d2da5ea56657\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-21T05:32:28.0059073+00:00\",\n  \"endTime\": \"2026-04-21T05:45:00.835958+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"3c10c973-53e0-49f2-94be-d2da5ea56657\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.35.1.cilium.base","result":"FAIL","duration":785903457082,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: PollUntilDone(ci-4676.0.0-n-0e35c4a698): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/1b4ada72-cd70-4fff-a9cc-79eaaa193ab5\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-21T05:32:27.9369277+00:00\",\n  \"endTime\": \"2026-04-21T05:45:00.4010617+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"1b4ada72-cd70-4fff-a9cc-79eaaa193ab5\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.locksmith.reboot","result":"FAIL","duration":1071469618867,"output":"        locksmith.go:141: failed to check rebooted machine: ssh unreachable or system not ready: context deadline exceeded\n"}],"result":"FAIL","platform":"azure","version":"4676.0.0"}
