{"tests":[{"name":"cl.ignition.symlink","result":"PASS","duration":96483239657,"output":"        cluster.go:152: + readlink /etc/localtime\n"},{"name":"cl.ignition.v2_1.vfat","result":"PASS","duration":97867453277,"output":""},{"name":"cl.ignition.v2.btrfsroot","result":"PASS","duration":99489757051,"output":""},{"name":"cl.ignition.v2.ext4root","result":"PASS","duration":99663992123,"output":""},{"name":"cl.network.nftables","result":"PASS","duration":126083584109,"output":"        cluster.go:152: + sudo nft --json list ruleset | jq '.nftables[] | select(.rule) | .rule.expr[0].match.right'\n"},{"name":"sysext.disable-docker","result":"PASS","duration":126364356944,"output":""},{"name":"cl.ignition.misc.empty","result":"PASS","duration":127677502670,"output":""},{"name":"cl.network.initramfs.second-boot","result":"PASS","duration":127755350689,"output":""},{"name":"coreos.locksmith.tls","result":"PASS","duration":131568574646,"output":""},{"name":"cl.osreset.ignition-rerun","result":"PASS","duration":137390563370,"output":""},{"name":"docker.selinux","result":"PASS","duration":158940889002,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: dbc3eadfebd7: Pulling fs layer\n        cluster.go:125: dbc3eadfebd7: Verifying Checksum\n        cluster.go:125: dbc3eadfebd7: Download complete\n        cluster.go:125: dbc3eadfebd7: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: sh: can't create /opt/hello: Permission denied\n"},{"name":"cl.toolbox.dnf-install","result":"PASS","duration":161217652747,"output":""},{"name":"cl.ignition.v1.once","result":"PASS","duration":168569814549,"output":""},{"name":"sysext.custom-docker.sysext","result":"PASS","duration":188192763054,"output":"        cluster.go:125: bash: line 1: docker: command not found\n        cluster.go:125: Cloning into 'sysext-bakery'...\n        cluster.go:125: Note: switching to '9850ffd5b2353f45a9b3bf4fb84f8138a149e3e7'.\n        cluster.go:125: \n        cluster.go:125: You are in 'detached HEAD' state. You can look around, make experimental\n        cluster.go:125: changes and commit them, and you can discard any commits you make in this\n        cluster.go:125: state without impacting any branches by switching back to a branch.\n        cluster.go:125: \n        cluster.go:125: If you want to create a new branch to retain commits you create, you may\n        cluster.go:125: do so (now or later) by using -c with the switch command. Example:\n        cluster.go:125: \n        cluster.go:125:   git switch -c \u003cnew-branch-name\u003e\n        cluster.go:125: \n        cluster.go:125: Or undo this operation with:\n        cluster.go:125: \n        cluster.go:125:   git switch -\n        cluster.go:125: \n        cluster.go:125: Turn off this advice by setting config variable advice.detachedHead to false\n        cluster.go:125: \n        cluster.go:125: HEAD is now at 9850ffd Merge pull request #31 from flatcar/t-lo/fix-docker-23-containerd-shim\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: dbc3eadfebd7: Pulling fs layer\n        cluster.go:125: dbc3eadfebd7: Verifying Checksum\n        cluster.go:125: dbc3eadfebd7: Download complete\n        cluster.go:125: dbc3eadfebd7: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: mke2fs 1.47.3 (8-Jul-2025)\n        cluster.go:125: resize2fs 1.47.3 (8-Jul-2025)\n"},{"name":"docker.enable-service.sysext","result":"PASS","duration":95552442304,"output":"        cluster.go:152: + systemctl is-enabled docker\n"},{"name":"cl.metadata.azure","result":"PASS","duration":96289409315,"output":""},{"name":"cl.update.reboot","result":"PASS","duration":222336503268,"output":""},{"name":"coreos.ignition.systemd.enable-service","result":"PASS","duration":124798604632,"output":""},{"name":"coreos.ignition.resource.local","result":"PASS","duration":230846412148,"output":""},{"name":"coreos.tls.fetch-urls","result":"PASS","duration":105141418233,"output":""},{"name":"bpf.ig/ig","result":"PASS","duration":110389828639,"output":"        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n        cluster.go:125: time=\"2026-04-28T23:32:12Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Error response from daemon: client version 1.51 is too new. Maximum supported API version is 1.50\"\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n        cluster.go:125: + docker info\n        cluster.go:125: + docker ps\n        cluster.go:125: + docker images\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n        cluster.go:125: + trap 'kill %%' ERR\n        cluster.go:125: + timeout 30 grep -F -m1 running...\n        cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n        cluster.go:125: + dig kinvolk.io\n        cluster.go:125: + dig flatcar.org\n        cluster.go:125: + dig stable.release.flatcar-linux.net\n        cluster.go:125: + kill %%\n        cluster.go:125: + wait\n        cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n        cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"bpf.ig","result":"PASS","duration":113243132997,"output":"    --- PASS: bpf.ig/ig (110.39s)\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --help\n            cluster.go:125: time=\"2026-04-28T23:32:12Z\" level=warning msg=\"Runtime enricher (docker): couldn't get current containers: listing containers with options {Size:false All:true Latest:false Since: Before: Limit:0 Filters:{fields:map[]}}: Error response from daemon: client version 1.51 is too new. Maximum supported API version is 1.50\"\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + sudo ig run trace_exec:v0.50.0 --host --filter proc.comm=docker,args~ps --output json --verbose\n            cluster.go:125: + docker info\n            cluster.go:125: + docker ps\n            cluster.go:125: + docker images\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.args == \"/usr/bin/docker\\u00a0ps\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0info\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.args == \"/usr/bin/docker\\u00a0images\"))' ig.json\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --help\n            cluster.go:125: + trap 'kill %%' ERR\n            cluster.go:125: + timeout 30 grep -F -m1 running...\n            cluster.go:125: + sudo ig run trace_dns:v0.50.0 --host --filter name=flatcar.org. --output json --verbose\n            cluster.go:125: + dig kinvolk.io\n            cluster.go:125: + dig flatcar.org\n            cluster.go:125: + dig stable.release.flatcar-linux.net\n            cluster.go:125: + kill %%\n            cluster.go:125: + wait\n            cluster.go:125: + jq -s -e '.[] | select(.name == \"flatcar.org.\")' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"kinvolk.io.\"))' ig.json\n            cluster.go:125: + jq -s -e 'isempty(.[] | select(.name == \"stable.release.flatcar-linux.net.\"))' ig.json\n"},{"name":"cl.swap_activation","result":"PASS","duration":125973802479,"output":""},{"name":"coreos.ignition.ssh.key","result":"PASS","duration":95665413504,"output":""},{"name":"cl.etcd-member.discovery","result":"PASS","duration":128590609712,"output":""},{"name":"coreos.ignition.once","result":"PASS","duration":129669437031,"output":""},{"name":"coreos.selinux.enforce","result":"PASS","duration":169644298005,"output":""},{"name":"cl.etcd-member.etcdctlv3","result":"PASS","duration":97497843999,"output":"        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-28T23:33:39.127563Z\",\"caller\":\"snapshot/v3_snapshot.go:65\",\"msg\":\"created temporary db file\",\"path\":\"/tmp/tmp.rGaTxWOfmv/snapshot.db.part\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-28T23:33:39.128416Z\",\"logger\":\"client\",\"caller\":\"v3@v3.5.18/maintenance.go:212\",\"msg\":\"opened snapshot stream; downloading\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-28T23:33:39.128443Z\",\"caller\":\"snapshot/v3_snapshot.go:73\",\"msg\":\"fetching snapshot\",\"endpoint\":\"127.0.0.1:2379\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-28T23:33:39.136332Z\",\"logger\":\"client\",\"caller\":\"v3@v3.5.18/maintenance.go:220\",\"msg\":\"completed snapshot read; closing\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-28T23:33:39.136779Z\",\"caller\":\"snapshot/v3_snapshot.go:88\",\"msg\":\"fetched snapshot\",\"endpoint\":\"127.0.0.1:2379\",\"size\":\"20 kB\",\"took\":\"now\"}\n        cluster.go:125: {\"level\":\"info\",\"ts\":\"2026-04-28T23:33:39.136956Z\",\"caller\":\"snapshot/v3_snapshot.go:97\",\"msg\":\"saved\",\"path\":\"/tmp/tmp.rGaTxWOfmv/snapshot.db\"}\n        cluster.go:125: Deprecated: Use `etcdutl snapshot status` instead.\n"},{"name":"coreos.update.badusr","result":"FAIL","duration":114989740831,"output":"        cluster.go:145: \"sudo cgpt prioritize /dev/disk/by-partlabel/USR-B\" failed: output , status ssh: handshake failed: EOF\n"},{"name":"coreos.selinux.boolean","result":"FAIL","duration":400852222,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-6f100375c1\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.overlay.cleanup","result":"FAIL","duration":242136514,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-86fbe269ca\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"docker.containerd-restart","result":"FAIL","duration":142211790178,"output":"        cluster.go:125: Unable to find image 'ghcr.io/flatcar/busybox:latest' locally\n        cluster.go:125: latest: Pulling from flatcar/busybox\n        cluster.go:125: dbc3eadfebd7: Pulling fs layer\n        cluster.go:125: dbc3eadfebd7: Verifying Checksum\n        cluster.go:125: dbc3eadfebd7: Download complete\n        cluster.go:125: dbc3eadfebd7: Pull complete\n        cluster.go:125: Digest: sha256:93e8234eb9ca92b9aae20fd73d6c9447ac3d1cc741c6e80c737f821dca582a0e\n        cluster.go:125: Status: Downloaded newer image for ghcr.io/flatcar/busybox:latest\n        cluster.go:145: \"systemctl show containerd -p SubState --value\" failed: output , status ssh: handshake failed: read tcp 10.200.1.7:45374-\u003e20.112.79.232:22: read: connection reset by peer\n"},{"name":"kubeadm.v1.35.1.cilium.base","result":"FAIL","duration":240383770953,"output":"        cluster.go:125: I0428 23:33:26.860121    3010 version.go:260] remote version is much newer: v1.36.0; falling back to: stable-1.35\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.35.4\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.35.4\n        kubeadm.go:197: unable to setup cluster: unable to run master script: wait: remote command exited without exit status or exit signal\n"},{"name":"sysext.disable-containerd","result":"FAIL","duration":226564820,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-82c9a399ed\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"docker.base","result":"FAIL","duration":222187479,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-dc8d8cf6a7\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.auth.verify","result":"FAIL","duration":576156003,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-e650204782\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.network.wireguard","result":"FAIL","duration":223205382,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-487fd7cb31\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"sysext.simple","result":"FAIL","duration":228601243,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-c4e8c1f7bd\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.internet","result":"FAIL","duration":232378456,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-69590069a3\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.cloudinit.script","result":"FAIL","duration":225149042,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-8e04f3bee2\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.luks","result":"FAIL","duration":205023369,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-f7ae1c6060\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.ext4root","result":"FAIL","duration":207787395,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-9bc819745b\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2.xfsroot","result":"FAIL","duration":388524643,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-896ee8ddcb\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2.users","result":"FAIL","duration":251105533,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-1cd2a3d341\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.cloudinit.multipart-mime","result":"FAIL","duration":229966011,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-0069b1e3b7\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"systemd.sysusers.gshadow","result":"FAIL","duration":247511408,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-1561192853\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.kargs","result":"FAIL","duration":214161772,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-fffd2c3ab5\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"docker.network-openbsd-nc","result":"FAIL","duration":263703165,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-10fcb147cb\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"systemd.journal.user","result":"FAIL","duration":234814571,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-0af72be73d\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2_1.swap","result":"FAIL","duration":216456161,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-d793c723ce\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.basic","result":"FAIL","duration":215849367,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-f4e84a498d\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2.noop","result":"FAIL","duration":228031793,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-34d256eff6\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.34.4.cilium.base","result":"FAIL","duration":18044647797,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: PollUntilDone(ci-4669.0.0-n-9c842415dc): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/f1932f55-bcb0-4057-9ae8-352350bf14ad\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:33:45.2498992+00:00\",\n  \"endTime\": \"2026-04-28T23:33:48.2509538+00:00\",\n  \"status\": \"Failed\",\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'KOLA-CLUSTER-IMAGE-E3B0059ED2' is in deprovisioning state and cannot perform this operation.  Target: '/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Compute/disks/ci-4669.0.0-n-9c842415dc_OsDisk_1_38e5e016aa914f1d846869fbbf7e4af8'.\"\n  },\n  \"name\": \"f1932f55-bcb0-4057-9ae8-352350bf14ad\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"docker.btrfs-storage","result":"FAIL","duration":210929806,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-e724e97c67\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.network.iptables","result":"FAIL","duration":228096337,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-67e977e5a9\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"docker.userns","result":"FAIL","duration":231457175,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-ee59970469\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.translation","result":"FAIL","duration":210309031,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-e8320a91dc\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.cloudinit.basic","result":"FAIL","duration":225584289,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-3ba9345705\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.etcd-member.v2-backup-restore","result":"FAIL","duration":617395000,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-e8cea9f988\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.groups","result":"FAIL","duration":321205658,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-89daa49924\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.ignition.security.tls","result":"FAIL","duration":212649259,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-4d707f41c3\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.flannel.vxlan","result":"FAIL","duration":869805177,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-4a28b39009\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.xfsroot","result":"FAIL","duration":234276248,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-aa2d3d38b6\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.btrfsroot","result":"FAIL","duration":215841448,"output":"        harness.go:646: Cluster failed starting machines: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-f3e57fb55c\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.35.1.calico.base","result":"FAIL","duration":2341461222,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-851d4bca36\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.33.8.cilium.base","result":"FAIL","duration":2216501250,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: creating public ip: PUT https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Network/publicIPAddresses/ip-5b893094a6\n--------------------------------------------------------------------------------\nRESPONSE 409: 409 Conflict\nERROR CODE: ResourceGroupBeingDeleted\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"ResourceGroupBeingDeleted\",\n    \"message\": \"The resource group 'kola-cluster-image-e3b0059ed2' is in deprovisioning state and cannot perform this operation.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.33.8.calico.base","result":"FAIL","duration":354503398619,"output":"        cluster.go:125: I0428 23:31:56.934065    2961 version.go:261] remote version is much newer: v1.36.0; falling back to: stable-1.33\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.33.11\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.0\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.5.24-0\n        cluster.go:125: I0428 23:32:09.639874    3191 version.go:261] remote version is much newer: v1.36.0; falling back to: stable-1.33\n        cluster.go:125: [init] Using Kubernetes version: v1.33.11\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669.0.0-n-933905b192\" could not be reached\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669.0.0-n-933905b192\": lookup ci-4669.0.0-n-933905b192 on 168.63.129.16:53: no such host\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [ci-4669.0.0-n-933905b192 kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 10.0.0.18]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.001521284s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.18:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 1.63473606s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 2.03808152s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 5.501543884s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669.0.0-n-933905b192 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669.0.0-n-933905b192 as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: p2d0lp.d3ddeiuyhmv1pvpd\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.18:6443 --token p2d0lp.d3ddeiuyhmv1pvpd \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:64e33374da1e24de8a8f868dbd179065040224da6e1d3cb9ab402a3dc00c80fa \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: installation.operator.tigera.io/default created\n        cluster.go:125: apiserver.operator.tigera.io/default created\n        cluster.go:125: goldmane.operator.tigera.io/default created\n        cluster.go:125: whisker.operator.tigera.io/default created\n        kubeadm.go:197: unable to setup cluster: unable to create worker node: waiting for machine to become active: GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Compute/virtualMachines/ci-4669.0.0-n-60423159a7\n--------------------------------------------------------------------------------\nRESPONSE 404: 404 Not Found\nERROR CODE: NotFound\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"NotFound\",\n    \"message\": \"The entity was not found in this Azure location.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.ignition.groups","result":"FAIL","duration":124308027355,"output":"        harness.go:646: Cluster failed starting machines: waiting for machine to become active: GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Compute/virtualMachines/ci-4669.0.0-n-12478cfc84\n--------------------------------------------------------------------------------\nRESPONSE 404: 404 Not Found\nERROR CODE: NotFound\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"NotFound\",\n    \"message\": \"The entity was not found in this Azure location.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"docker.lib-coreos-dockerd-compat","result":"FAIL","duration":124329739155,"output":"        harness.go:646: Cluster failed starting machines: waiting for machine to become active: GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Compute/virtualMachines/ci-4669.0.0-n-08dfadb23c\n--------------------------------------------------------------------------------\nRESPONSE 404: 404 Not Found\nERROR CODE: NotFound\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"NotFound\",\n    \"message\": \"The entity was not found in this Azure location.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.34.4.flannel.base","result":"FAIL","duration":359270877714,"output":"        cluster.go:125: I0428 23:32:09.476088    2739 version.go:260] remote version is much newer: v1.36.0; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0428 23:32:20.767620    3085 version.go:260] remote version is much newer: v1.36.0; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669.0.0-n-dedb59046c\" could not be reached\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669.0.0-n-dedb59046c\": lookup ci-4669.0.0-n-dedb59046c on 168.63.129.16:53: no such host\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [ci-4669.0.0-n-dedb59046c kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 10.0.0.24]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.001606565s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.24:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 2.00498546s\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 2.405876537s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 4.501973837s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669.0.0-n-dedb59046c as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669.0.0-n-dedb59046c as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: c2nktx.5irfer599dzp0zkp\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.24:6443 --token c2nktx.5irfer599dzp0zkp \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:18f253d148de496d3862f1528c37d2c250e53964450a90b0ae93e3ea1ea49f08 \n        cluster.go:125: namespace/kube-flannel created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/flannel created\n        cluster.go:125: serviceaccount/flannel created\n        cluster.go:125: configmap/kube-flannel-cfg created\n        cluster.go:125: daemonset.apps/kube-flannel-ds created\n        kubeadm.go:197: unable to setup cluster: unable to create worker node: waiting for machine to become active: GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/resourceGroups/kola-cluster-image-e3b0059ed2/providers/Microsoft.Compute/virtualMachines/ci-4669.0.0-n-1b39428bca\n--------------------------------------------------------------------------------\nRESPONSE 404: 404 Not Found\nERROR CODE: NotFound\n--------------------------------------------------------------------------------\n{\n  \"error\": {\n    \"code\": \"NotFound\",\n    \"message\": \"The entity was not found in this Azure location.\"\n  }\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.33.8.flannel.base","result":"FAIL","duration":176520581684,"output":"        kubeadm.go:197: unable to setup cluster: unable to create master node: PollUntilDone(ci-4669.0.0-n-819d8119b3): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/1f60b143-9c72-4905-9863-ed2392c1aec9\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:33:33.6778912+00:00\",\n  \"endTime\": \"2026-04-28T23:34:07.4004577+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"1f60b143-9c72-4905-9863-ed2392c1aec9\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.35.1.flannel.base","result":"FAIL","duration":114115763112,"output":"        kubeadm.go:197: unable to setup cluster: unable to create etcd node: PollUntilDone(ci-4669.0.0-n-63f1f1835e): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/4b830c5c-4629-4eeb-bb67-e0433da9c531\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:32:33.7210883+00:00\",\n  \"endTime\": \"2026-04-28T23:34:21.159013+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"4b830c5c-4629-4eeb-bb67-e0433da9c531\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.noop","result":"FAIL","duration":112742323745,"output":"        harness.go:646: Cluster failed starting machines: PollUntilDone(ci-4669.0.0-n-bba253b3f4): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/1f51448a-fa32-471e-9273-a5b1855136be\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:32:35.5499661+00:00\",\n  \"endTime\": \"2026-04-28T23:34:07.4004577+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"1f51448a-fa32-471e-9273-a5b1855136be\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"kubeadm.v1.34.4.calico.base","result":"FAIL","duration":368535596005,"output":"        cluster.go:125: I0428 23:32:06.098233    2963 version.go:260] remote version is much newer: v1.36.0; falling back to: stable-1.34\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-apiserver:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-controller-manager:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-scheduler:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/kube-proxy:v1.34.7\n        cluster.go:125: [config/images] Pulled registry.k8s.io/coredns/coredns:v1.12.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/pause:3.10.1\n        cluster.go:125: [config/images] Pulled registry.k8s.io/etcd:3.6.5-0\n        cluster.go:125: I0428 23:33:02.352177    3222 version.go:260] remote version is much newer: v1.36.0; falling back to: stable-1.34\n        cluster.go:125: [init] Using Kubernetes version: v1.34.7\n        cluster.go:125: [preflight] Running pre-flight checks\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669.0.0-n-55a7652d44\" could not be reached\n        cluster.go:125: \t[WARNING Hostname]: hostname \"ci-4669.0.0-n-55a7652d44\": lookup ci-4669.0.0-n-55a7652d44 on 168.63.129.16:53: no such host\n        cluster.go:125: \t[WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'\n        cluster.go:125: [preflight] Pulling images required for setting up a Kubernetes cluster\n        cluster.go:125: [preflight] This might take a minute or two, depending on the speed of your internet connection\n        cluster.go:125: [preflight] You can also perform this action beforehand using 'kubeadm config images pull'\n        cluster.go:125: [certs] Using certificateDir folder \"/etc/kubernetes/pki\"\n        cluster.go:125: [certs] Generating \"ca\" certificate and key\n        cluster.go:125: [certs] Generating \"apiserver\" certificate and key\n        cluster.go:125: [certs] apiserver serving cert is signed for DNS names [ci-4669.0.0-n-55a7652d44 kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 10.0.0.22]\n        cluster.go:125: [certs] Generating \"apiserver-kubelet-client\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-ca\" certificate and key\n        cluster.go:125: [certs] Generating \"front-proxy-client\" certificate and key\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/ca certificate authority generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/server certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/peer certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping etcd/healthcheck-client certificate generation\n        cluster.go:125: [certs] External etcd mode: Skipping apiserver-etcd-client certificate generation\n        cluster.go:125: [certs] Generating \"sa\" key and public key\n        cluster.go:125: [kubeconfig] Using kubeconfig folder \"/etc/kubernetes\"\n        cluster.go:125: [kubeconfig] Writing \"admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"super-admin.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"kubelet.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"controller-manager.conf\" kubeconfig file\n        cluster.go:125: [kubeconfig] Writing \"scheduler.conf\" kubeconfig file\n        cluster.go:125: [control-plane] Using manifest folder \"/etc/kubernetes/manifests\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-apiserver\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-controller-manager\"\n        cluster.go:125: [control-plane] Creating static Pod manifest for \"kube-scheduler\"\n        cluster.go:125: [kubelet-start] Writing kubelet environment file with flags to file \"/var/lib/kubelet/kubeadm-flags.env\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/instance-config.yaml\"\n        cluster.go:125: [patches] Applied patch of type \"application/strategic-merge-patch+json\" to target \"kubeletconfiguration\"\n        cluster.go:125: [kubelet-start] Writing kubelet configuration to file \"/var/lib/kubelet/config.yaml\"\n        cluster.go:125: [kubelet-start] Starting the kubelet\n        cluster.go:125: [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory \"/etc/kubernetes/manifests\"\n        cluster.go:125: [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s\n        cluster.go:125: [kubelet-check] The kubelet is healthy after 1.015745954s\n        cluster.go:125: [control-plane-check] Waiting for healthy control plane components. This can take up to 30m0s\n        cluster.go:125: [control-plane-check] Checking kube-apiserver at https://10.0.0.22:6443/livez\n        cluster.go:125: [control-plane-check] Checking kube-controller-manager at https://127.0.0.1:10257/healthz\n        cluster.go:125: [control-plane-check] Checking kube-scheduler at https://127.0.0.1:10259/livez\n        cluster.go:125: [control-plane-check] kube-scheduler is healthy after 4.005201245s\n        cluster.go:125: [control-plane-check] kube-apiserver is healthy after 6.501848803s\n        cluster.go:125: [control-plane-check] kube-controller-manager is healthy after 7.004132687s\n        cluster.go:125: [upload-config] Storing the configuration used in ConfigMap \"kubeadm-config\" in the \"kube-system\" Namespace\n        cluster.go:125: [kubelet] Creating a ConfigMap \"kubelet-config\" in namespace kube-system with the configuration for the kubelets in the cluster\n        cluster.go:125: [upload-certs] Skipping phase. Please see --upload-certs\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669.0.0-n-55a7652d44 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]\n        cluster.go:125: [mark-control-plane] Marking the node ci-4669.0.0-n-55a7652d44 as control-plane by adding the taints [node-role.kubernetes.io/control-plane:NoSchedule]\n        cluster.go:125: [bootstrap-token] Using token: rvotw5.yp1wac6oxoyq4xsq\n        cluster.go:125: [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token\n        cluster.go:125: [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster\n        cluster.go:125: [bootstrap-token] Creating the \"cluster-info\" ConfigMap in the \"kube-public\" namespace\n        cluster.go:125: [kubelet-finalize] Updating \"/etc/kubernetes/kubelet.conf\" to point to a rotatable kubelet client certificate and key\n        cluster.go:125: [addons] Applied essential addon: CoreDNS\n        cluster.go:125: [addons] Applied essential addon: kube-proxy\n        cluster.go:125: \n        cluster.go:125: Your Kubernetes control-plane has initialized successfully!\n        cluster.go:125: \n        cluster.go:125: To start using your cluster, you need to run the following as a regular user:\n        cluster.go:125: \n        cluster.go:125:   mkdir -p $HOME/.kube\n        cluster.go:125:   sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config\n        cluster.go:125:   sudo chown $(id -u):$(id -g) $HOME/.kube/config\n        cluster.go:125: \n        cluster.go:125: Alternatively, if you are the root user, you can run:\n        cluster.go:125: \n        cluster.go:125:   export KUBECONFIG=/etc/kubernetes/admin.conf\n        cluster.go:125: \n        cluster.go:125: You should now deploy a pod network to the cluster.\n        cluster.go:125: Run \"kubectl apply -f [podnetwork].yaml\" with one of the options listed at:\n        cluster.go:125:   https://kubernetes.io/docs/concepts/cluster-administration/addons/\n        cluster.go:125: \n        cluster.go:125: Then you can join any number of worker nodes by running the following on each as root:\n        cluster.go:125: \n        cluster.go:125: kubeadm join 10.0.0.22:6443 --token rvotw5.yp1wac6oxoyq4xsq \\\n        cluster.go:125: \t--discovery-token-ca-cert-hash sha256:2a84603233d4846be6a88c2426e3cdb7e949d91c5f4820bee92c7214e8d27e5a \n        cluster.go:125: namespace/tigera-operator created\n        cluster.go:125: serviceaccount/tigera-operator created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: clusterrole.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: clusterrolebinding.rbac.authorization.k8s.io/tigera-operator created\n        cluster.go:125: rolebinding.rbac.authorization.k8s.io/tigera-operator-secrets created\n        cluster.go:125: deployment.apps/tigera-operator created\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/installations.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: customresourcedefinition.apiextensions.k8s.io/apiservers.operator.tigera.io condition met\n        cluster.go:125: installation.operator.tigera.io/default created\n        cluster.go:125: apiserver.operator.tigera.io/default created\n        cluster.go:125: goldmane.operator.tigera.io/default created\n        cluster.go:125: whisker.operator.tigera.io/default created\n        kubeadm.go:197: unable to setup cluster: unable to create worker node: PollUntilDone(ci-4669.0.0-n-ca81187d08): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/a2b24844-99b9-4371-8982-011db0b1b224\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:33:37.9351702+00:00\",\n  \"endTime\": \"2026-04-28T23:34:07.5230526+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"a2b24844-99b9-4371-8982-011db0b1b224\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.users.shells","result":"FAIL","duration":82740751169,"output":"        harness.go:646: Cluster failed starting machines: PollUntilDone(ci-4669.0.0-n-5d61ddf0c8): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/18b3095b-9d4b-4d16-b04e-fdf4454146a8\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:33:11.3152737+00:00\",\n  \"endTime\": \"2026-04-28T23:34:13.2139627+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"18b3095b-9d4b-4d16-b04e-fdf4454146a8\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v2_1.ext4checkexisting","result":"FAIL","duration":178504632572,"output":"        filesystem.go:415: Couldn't start machine: PollUntilDone(ci-4669.0.0-n-724245ddc3): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/81c13958-17f3-409c-96c3-698da10137a7\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:33:09.1836221+00:00\",\n  \"endTime\": \"2026-04-28T23:34:07.1387017+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"81c13958-17f3-409c-96c3-698da10137a7\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.ignition.resource.remote","result":"FAIL","duration":112190159674,"output":"        harness.go:646: Cluster failed starting machines: PollUntilDone(ci-4669.0.0-n-56fb5ce248): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/168e3252-13e1-425a-942b-773bcff12c3f\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:32:47.5095383+00:00\",\n  \"endTime\": \"2026-04-28T23:34:07.2847039+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"168e3252-13e1-425a-942b-773bcff12c3f\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.verity","result":"FAIL","duration":80878417459,"output":"        harness.go:646: Cluster failed starting machines: PollUntilDone(ci-4669.0.0-n-37834d8422): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/0f79cadd-5497-4801-9ae8-31407a95634b\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:33:18.7720474+00:00\",\n  \"endTime\": \"2026-04-28T23:34:07.2847039+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"0f79cadd-5497-4801-9ae8-31407a95634b\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"cl.ignition.v1.users","result":"FAIL","duration":787702184341,"output":"        harness.go:646: Cluster failed starting machines: PollUntilDone(ci-4669.0.0-n-64ad832ff3): GET https://management.azure.com/subscriptions/0e46bd28-a80f-4d3a-8200-d9eb8d80cb2e/providers/Microsoft.Compute/locations/westus2/operations/7501c794-3944-4b8c-9850-aea188a14ea7\n--------------------------------------------------------------------------------\nRESPONSE 200: 200 OK\nERROR CODE: OperationPreempted\n--------------------------------------------------------------------------------\n{\n  \"startTime\": \"2026-04-28T23:33:14.5641887+00:00\",\n  \"endTime\": \"2026-04-28T23:45:59.6589709+00:00\",\n  \"status\": \"Canceled\",\n  \"error\": {\n    \"code\": \"OperationPreempted\",\n    \"message\": \"Operation execution has been preempted by a more recent operation.\"\n  },\n  \"name\": \"7501c794-3944-4b8c-9850-aea188a14ea7\"\n}\n--------------------------------------------------------------------------------\n"},{"name":"coreos.locksmith.reboot","result":"FAIL","duration":1223239817751,"output":"        locksmith.go:141: failed to check rebooted machine: ssh unreachable or system not ready: context deadline exceeded\n"},{"name":"cl.locksmith.cluster","result":"FAIL","duration":1229616499053,"output":"        locksmith.go:184: [0] ssh unreachable or system not ready: context deadline exceeded [1] ssh unreachable or system not ready: context deadline exceeded [2] ssh unreachable or system not ready: context deadline exceeded\n"},{"name":"cl.update.badverity","result":"FAIL","duration":5201702051529,"output":"        update.go:168: ssh unreachable or system not ready: failure checking if machine is running: systemctl is-system-running returned stdout: \"\", stderr: \"\", err: dial tcp 20.230.175.226:22: i/o timeout, systemctl list-jobs returned stdout: \"\", stderr: \"\", err: dial tcp 20.230.175.226:22: i/o timeout\n"}],"result":"FAIL","platform":"azure","version":"4669.0.0"}
