Apr 30 00:36:13.305801 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Apr 30 00:36:13.305824 kernel: Linux version 6.6.88-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Tue Apr 29 23:08:45 -00 2025 Apr 30 00:36:13.305832 kernel: KASLR enabled Apr 30 00:36:13.305838 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Apr 30 00:36:13.305845 kernel: printk: bootconsole [pl11] enabled Apr 30 00:36:13.305850 kernel: efi: EFI v2.7 by EDK II Apr 30 00:36:13.305858 kernel: efi: ACPI 2.0=0x3fd5f018 SMBIOS=0x3e580000 SMBIOS 3.0=0x3e560000 MEMATTR=0x3ead8b98 RNG=0x3fd5f998 MEMRESERVE=0x3e44ee18 Apr 30 00:36:13.305864 kernel: random: crng init done Apr 30 00:36:13.305870 kernel: ACPI: Early table checksum verification disabled Apr 30 00:36:13.305875 kernel: ACPI: RSDP 0x000000003FD5F018 000024 (v02 VRTUAL) Apr 30 00:36:13.305881 kernel: ACPI: XSDT 0x000000003FD5FF18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305887 kernel: ACPI: FACP 0x000000003FD5FC18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305894 kernel: ACPI: DSDT 0x000000003FD41018 01DFCD (v02 MSFTVM DSDT01 00000001 INTL 20230628) Apr 30 00:36:13.305900 kernel: ACPI: DBG2 0x000000003FD5FB18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305907 kernel: ACPI: GTDT 0x000000003FD5FD98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305914 kernel: ACPI: OEM0 0x000000003FD5F098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305920 kernel: ACPI: SPCR 0x000000003FD5FA98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305928 kernel: ACPI: APIC 0x000000003FD5F818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305934 kernel: ACPI: SRAT 0x000000003FD5F198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305940 kernel: ACPI: PPTT 0x000000003FD5F418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Apr 30 00:36:13.305947 kernel: ACPI: BGRT 0x000000003FD5FE98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305953 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Apr 30 00:36:13.305959 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] Apr 30 00:36:13.305965 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] Apr 30 00:36:13.305972 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] Apr 30 00:36:13.305978 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] Apr 30 00:36:13.305984 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] Apr 30 00:36:13.305990 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] Apr 30 00:36:13.305998 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] Apr 30 00:36:13.306004 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] Apr 30 00:36:13.306010 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] Apr 30 00:36:13.306016 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] Apr 30 00:36:13.306023 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] Apr 30 00:36:13.306029 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] Apr 30 00:36:13.306035 kernel: NUMA: NODE_DATA [mem 0x1bf7ee800-0x1bf7f3fff] Apr 30 00:36:13.306041 kernel: Zone ranges: Apr 30 00:36:13.306048 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Apr 30 00:36:13.306054 kernel: DMA32 empty Apr 30 00:36:13.306060 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Apr 30 00:36:13.306066 kernel: Movable zone start for each node Apr 30 00:36:13.306077 kernel: Early memory node ranges Apr 30 00:36:13.306083 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Apr 30 00:36:13.306090 kernel: node 0: [mem 0x0000000000824000-0x000000003e54ffff] Apr 30 00:36:13.306096 kernel: node 0: [mem 0x000000003e550000-0x000000003e87ffff] Apr 30 00:36:13.306103 kernel: node 0: [mem 0x000000003e880000-0x000000003fc7ffff] Apr 30 00:36:13.306110 kernel: node 0: [mem 0x000000003fc80000-0x000000003fcfffff] Apr 30 00:36:13.306117 kernel: node 0: [mem 0x000000003fd00000-0x000000003fffffff] Apr 30 00:36:13.306124 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Apr 30 00:36:13.306130 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Apr 30 00:36:13.306137 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Apr 30 00:36:13.306144 kernel: psci: probing for conduit method from ACPI. Apr 30 00:36:13.306150 kernel: psci: PSCIv1.1 detected in firmware. Apr 30 00:36:13.306157 kernel: psci: Using standard PSCI v0.2 function IDs Apr 30 00:36:13.306163 kernel: psci: MIGRATE_INFO_TYPE not supported. Apr 30 00:36:13.306170 kernel: psci: SMC Calling Convention v1.4 Apr 30 00:36:13.306176 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Apr 30 00:36:13.306183 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Apr 30 00:36:13.306191 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Apr 30 00:36:13.306198 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Apr 30 00:36:13.306205 kernel: pcpu-alloc: [0] 0 [0] 1 Apr 30 00:36:13.306211 kernel: Detected PIPT I-cache on CPU0 Apr 30 00:36:13.306218 kernel: CPU features: detected: GIC system register CPU interface Apr 30 00:36:13.306225 kernel: CPU features: detected: Hardware dirty bit management Apr 30 00:36:13.306231 kernel: CPU features: detected: Spectre-BHB Apr 30 00:36:13.306238 kernel: CPU features: kernel page table isolation forced ON by KASLR Apr 30 00:36:13.306245 kernel: CPU features: detected: Kernel page table isolation (KPTI) Apr 30 00:36:13.306251 kernel: CPU features: detected: ARM erratum 1418040 Apr 30 00:36:13.306258 kernel: CPU features: detected: ARM erratum 1542419 (kernel portion) Apr 30 00:36:13.306266 kernel: CPU features: detected: SSBS not fully self-synchronizing Apr 30 00:36:13.306272 kernel: alternatives: applying boot alternatives Apr 30 00:36:13.306280 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=2f2ec97241771b99b21726307071be4f8c5924f9157dc58cd38c4fcfbe71412a Apr 30 00:36:13.306288 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Apr 30 00:36:13.306294 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 30 00:36:13.306301 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 30 00:36:13.306307 kernel: Fallback order for Node 0: 0 Apr 30 00:36:13.306314 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1032156 Apr 30 00:36:13.306321 kernel: Policy zone: Normal Apr 30 00:36:13.306327 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 30 00:36:13.306334 kernel: software IO TLB: area num 2. Apr 30 00:36:13.306342 kernel: software IO TLB: mapped [mem 0x000000003a44e000-0x000000003e44e000] (64MB) Apr 30 00:36:13.306349 kernel: Memory: 3982688K/4194160K available (10240K kernel code, 2186K rwdata, 8104K rodata, 39424K init, 897K bss, 211472K reserved, 0K cma-reserved) Apr 30 00:36:13.306356 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Apr 30 00:36:13.306362 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 30 00:36:13.306370 kernel: rcu: RCU event tracing is enabled. Apr 30 00:36:13.306377 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Apr 30 00:36:13.306384 kernel: Trampoline variant of Tasks RCU enabled. Apr 30 00:36:13.306390 kernel: Tracing variant of Tasks RCU enabled. Apr 30 00:36:13.306397 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 30 00:36:13.306404 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Apr 30 00:36:13.306411 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Apr 30 00:36:13.306418 kernel: GICv3: 960 SPIs implemented Apr 30 00:36:13.306425 kernel: GICv3: 0 Extended SPIs implemented Apr 30 00:36:13.306432 kernel: Root IRQ handler: gic_handle_irq Apr 30 00:36:13.306438 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Apr 30 00:36:13.306445 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Apr 30 00:36:13.306452 kernel: ITS: No ITS available, not enabling LPIs Apr 30 00:36:13.306458 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 30 00:36:13.306465 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 00:36:13.306472 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Apr 30 00:36:13.306478 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Apr 30 00:36:13.306485 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Apr 30 00:36:13.306493 kernel: Console: colour dummy device 80x25 Apr 30 00:36:13.306501 kernel: printk: console [tty1] enabled Apr 30 00:36:13.306507 kernel: ACPI: Core revision 20230628 Apr 30 00:36:13.306514 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Apr 30 00:36:13.306522 kernel: pid_max: default: 32768 minimum: 301 Apr 30 00:36:13.306529 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Apr 30 00:36:13.306535 kernel: landlock: Up and running. Apr 30 00:36:13.306542 kernel: SELinux: Initializing. Apr 30 00:36:13.306549 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 30 00:36:13.306556 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 30 00:36:13.306564 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 30 00:36:13.306571 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 30 00:36:13.306578 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3a8030, hints 0xe, misc 0x31e1 Apr 30 00:36:13.306585 kernel: Hyper-V: Host Build 10.0.22477.1619-1-0 Apr 30 00:36:13.306591 kernel: Hyper-V: enabling crash_kexec_post_notifiers Apr 30 00:36:13.306598 kernel: rcu: Hierarchical SRCU implementation. Apr 30 00:36:13.306605 kernel: rcu: Max phase no-delay instances is 400. Apr 30 00:36:13.306618 kernel: Remapping and enabling EFI services. Apr 30 00:36:13.306626 kernel: smp: Bringing up secondary CPUs ... Apr 30 00:36:13.306647 kernel: Detected PIPT I-cache on CPU1 Apr 30 00:36:13.306655 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Apr 30 00:36:13.306664 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 00:36:13.306671 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Apr 30 00:36:13.306678 kernel: smp: Brought up 1 node, 2 CPUs Apr 30 00:36:13.306686 kernel: SMP: Total of 2 processors activated. Apr 30 00:36:13.306693 kernel: CPU features: detected: 32-bit EL0 Support Apr 30 00:36:13.306701 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Apr 30 00:36:13.306709 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Apr 30 00:36:13.306716 kernel: CPU features: detected: CRC32 instructions Apr 30 00:36:13.306723 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Apr 30 00:36:13.306730 kernel: CPU features: detected: LSE atomic instructions Apr 30 00:36:13.306738 kernel: CPU features: detected: Privileged Access Never Apr 30 00:36:13.306745 kernel: CPU: All CPU(s) started at EL1 Apr 30 00:36:13.306752 kernel: alternatives: applying system-wide alternatives Apr 30 00:36:13.306759 kernel: devtmpfs: initialized Apr 30 00:36:13.306768 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 30 00:36:13.306775 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Apr 30 00:36:13.306783 kernel: pinctrl core: initialized pinctrl subsystem Apr 30 00:36:13.306790 kernel: SMBIOS 3.1.0 present. Apr 30 00:36:13.306797 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 09/28/2024 Apr 30 00:36:13.306804 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 30 00:36:13.306812 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Apr 30 00:36:13.306819 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Apr 30 00:36:13.306826 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Apr 30 00:36:13.306835 kernel: audit: initializing netlink subsys (disabled) Apr 30 00:36:13.306842 kernel: audit: type=2000 audit(0.047:1): state=initialized audit_enabled=0 res=1 Apr 30 00:36:13.306850 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 30 00:36:13.306857 kernel: cpuidle: using governor menu Apr 30 00:36:13.306864 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Apr 30 00:36:13.306871 kernel: ASID allocator initialised with 32768 entries Apr 30 00:36:13.306879 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 30 00:36:13.306886 kernel: Serial: AMBA PL011 UART driver Apr 30 00:36:13.306893 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Apr 30 00:36:13.306902 kernel: Modules: 0 pages in range for non-PLT usage Apr 30 00:36:13.306909 kernel: Modules: 509024 pages in range for PLT usage Apr 30 00:36:13.306917 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 30 00:36:13.306924 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Apr 30 00:36:13.306931 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Apr 30 00:36:13.306938 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Apr 30 00:36:13.306945 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 30 00:36:13.306953 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Apr 30 00:36:13.306960 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Apr 30 00:36:13.306968 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Apr 30 00:36:13.306976 kernel: ACPI: Added _OSI(Module Device) Apr 30 00:36:13.306983 kernel: ACPI: Added _OSI(Processor Device) Apr 30 00:36:13.306990 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Apr 30 00:36:13.306997 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 30 00:36:13.307004 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 30 00:36:13.307011 kernel: ACPI: Interpreter enabled Apr 30 00:36:13.307018 kernel: ACPI: Using GIC for interrupt routing Apr 30 00:36:13.307026 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Apr 30 00:36:13.307034 kernel: printk: console [ttyAMA0] enabled Apr 30 00:36:13.307041 kernel: printk: bootconsole [pl11] disabled Apr 30 00:36:13.307049 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Apr 30 00:36:13.307056 kernel: iommu: Default domain type: Translated Apr 30 00:36:13.307063 kernel: iommu: DMA domain TLB invalidation policy: strict mode Apr 30 00:36:13.307070 kernel: efivars: Registered efivars operations Apr 30 00:36:13.307078 kernel: vgaarb: loaded Apr 30 00:36:13.307085 kernel: clocksource: Switched to clocksource arch_sys_counter Apr 30 00:36:13.307092 kernel: VFS: Disk quotas dquot_6.6.0 Apr 30 00:36:13.307101 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 30 00:36:13.307108 kernel: pnp: PnP ACPI init Apr 30 00:36:13.307115 kernel: pnp: PnP ACPI: found 0 devices Apr 30 00:36:13.307122 kernel: NET: Registered PF_INET protocol family Apr 30 00:36:13.307129 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 30 00:36:13.307137 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 30 00:36:13.307144 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 30 00:36:13.307151 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 30 00:36:13.307159 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 30 00:36:13.307167 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 30 00:36:13.307175 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 30 00:36:13.307182 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 30 00:36:13.307189 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 30 00:36:13.307196 kernel: PCI: CLS 0 bytes, default 64 Apr 30 00:36:13.307204 kernel: kvm [1]: HYP mode not available Apr 30 00:36:13.307211 kernel: Initialise system trusted keyrings Apr 30 00:36:13.307218 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 30 00:36:13.307225 kernel: Key type asymmetric registered Apr 30 00:36:13.307233 kernel: Asymmetric key parser 'x509' registered Apr 30 00:36:13.307241 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Apr 30 00:36:13.307248 kernel: io scheduler mq-deadline registered Apr 30 00:36:13.307255 kernel: io scheduler kyber registered Apr 30 00:36:13.307262 kernel: io scheduler bfq registered Apr 30 00:36:13.307270 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 30 00:36:13.307277 kernel: thunder_xcv, ver 1.0 Apr 30 00:36:13.307284 kernel: thunder_bgx, ver 1.0 Apr 30 00:36:13.307291 kernel: nicpf, ver 1.0 Apr 30 00:36:13.307298 kernel: nicvf, ver 1.0 Apr 30 00:36:13.307449 kernel: rtc-efi rtc-efi.0: registered as rtc0 Apr 30 00:36:13.307525 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-04-30T00:36:12 UTC (1745973372) Apr 30 00:36:13.307535 kernel: efifb: probing for efifb Apr 30 00:36:13.307543 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Apr 30 00:36:13.307550 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Apr 30 00:36:13.307557 kernel: efifb: scrolling: redraw Apr 30 00:36:13.307564 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Apr 30 00:36:13.307574 kernel: Console: switching to colour frame buffer device 128x48 Apr 30 00:36:13.307582 kernel: fb0: EFI VGA frame buffer device Apr 30 00:36:13.307589 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Apr 30 00:36:13.307596 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 30 00:36:13.307603 kernel: No ACPI PMU IRQ for CPU0 Apr 30 00:36:13.307611 kernel: No ACPI PMU IRQ for CPU1 Apr 30 00:36:13.307618 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 1 counters available Apr 30 00:36:13.307625 kernel: watchdog: Delayed init of the lockup detector failed: -19 Apr 30 00:36:13.307644 kernel: watchdog: Hard watchdog permanently disabled Apr 30 00:36:13.307653 kernel: NET: Registered PF_INET6 protocol family Apr 30 00:36:13.307660 kernel: Segment Routing with IPv6 Apr 30 00:36:13.307667 kernel: In-situ OAM (IOAM) with IPv6 Apr 30 00:36:13.307675 kernel: NET: Registered PF_PACKET protocol family Apr 30 00:36:13.307682 kernel: Key type dns_resolver registered Apr 30 00:36:13.307689 kernel: registered taskstats version 1 Apr 30 00:36:13.307697 kernel: Loading compiled-in X.509 certificates Apr 30 00:36:13.307704 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.88-flatcar: e2b28159d3a83b6f5d5db45519e470b1b834e378' Apr 30 00:36:13.307711 kernel: Key type .fscrypt registered Apr 30 00:36:13.307720 kernel: Key type fscrypt-provisioning registered Apr 30 00:36:13.307727 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 30 00:36:13.307735 kernel: ima: Allocated hash algorithm: sha1 Apr 30 00:36:13.307742 kernel: ima: No architecture policies found Apr 30 00:36:13.307749 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Apr 30 00:36:13.307756 kernel: clk: Disabling unused clocks Apr 30 00:36:13.307763 kernel: Freeing unused kernel memory: 39424K Apr 30 00:36:13.307771 kernel: Run /init as init process Apr 30 00:36:13.307778 kernel: with arguments: Apr 30 00:36:13.307786 kernel: /init Apr 30 00:36:13.307794 kernel: with environment: Apr 30 00:36:13.307801 kernel: HOME=/ Apr 30 00:36:13.307808 kernel: TERM=linux Apr 30 00:36:13.307814 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Apr 30 00:36:13.307824 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 30 00:36:13.307834 systemd[1]: Detected virtualization microsoft. Apr 30 00:36:13.307842 systemd[1]: Detected architecture arm64. Apr 30 00:36:13.307851 systemd[1]: Running in initrd. Apr 30 00:36:13.307859 systemd[1]: No hostname configured, using default hostname. Apr 30 00:36:13.307866 systemd[1]: Hostname set to . Apr 30 00:36:13.307874 systemd[1]: Initializing machine ID from random generator. Apr 30 00:36:13.307882 systemd[1]: Queued start job for default target initrd.target. Apr 30 00:36:13.307890 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 00:36:13.307898 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 00:36:13.307906 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 30 00:36:13.307916 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 30 00:36:13.307924 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 30 00:36:13.307932 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 30 00:36:13.307941 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 30 00:36:13.307950 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 30 00:36:13.307957 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 00:36:13.307965 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 30 00:36:13.307974 systemd[1]: Reached target paths.target - Path Units. Apr 30 00:36:13.307982 systemd[1]: Reached target slices.target - Slice Units. Apr 30 00:36:13.307990 systemd[1]: Reached target swap.target - Swaps. Apr 30 00:36:13.307998 systemd[1]: Reached target timers.target - Timer Units. Apr 30 00:36:13.308006 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 30 00:36:13.308014 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 30 00:36:13.308021 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 30 00:36:13.308029 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 30 00:36:13.308039 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 30 00:36:13.308047 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 30 00:36:13.308055 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 00:36:13.308063 systemd[1]: Reached target sockets.target - Socket Units. Apr 30 00:36:13.308071 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 30 00:36:13.308078 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 30 00:36:13.308086 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 30 00:36:13.308094 systemd[1]: Starting systemd-fsck-usr.service... Apr 30 00:36:13.308102 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 30 00:36:13.308112 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 30 00:36:13.308138 systemd-journald[217]: Collecting audit messages is disabled. Apr 30 00:36:13.308158 systemd-journald[217]: Journal started Apr 30 00:36:13.308178 systemd-journald[217]: Runtime Journal (/run/log/journal/0998849b57cc425bbb6a442302b8b3b6) is 8.0M, max 78.5M, 70.5M free. Apr 30 00:36:13.319722 systemd-modules-load[218]: Inserted module 'overlay' Apr 30 00:36:13.329969 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 00:36:13.345655 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 30 00:36:13.345703 systemd[1]: Started systemd-journald.service - Journal Service. Apr 30 00:36:13.354671 kernel: Bridge firewalling registered Apr 30 00:36:13.354816 systemd-modules-load[218]: Inserted module 'br_netfilter' Apr 30 00:36:13.360870 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 30 00:36:13.382788 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 00:36:13.389740 systemd[1]: Finished systemd-fsck-usr.service. Apr 30 00:36:13.400084 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 30 00:36:13.409854 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:13.430931 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 00:36:13.439868 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 30 00:36:13.455849 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 30 00:36:13.483866 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 30 00:36:13.493105 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 00:36:13.507698 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 30 00:36:13.525815 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 30 00:36:13.550397 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 00:36:13.566132 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 30 00:36:13.578002 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 30 00:36:13.589221 dracut-cmdline[251]: dracut-dracut-053 Apr 30 00:36:13.593577 dracut-cmdline[251]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=2f2ec97241771b99b21726307071be4f8c5924f9157dc58cd38c4fcfbe71412a Apr 30 00:36:13.624091 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 30 00:36:13.645218 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 00:36:13.657584 systemd-resolved[260]: Positive Trust Anchors: Apr 30 00:36:13.657593 systemd-resolved[260]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 30 00:36:13.657624 systemd-resolved[260]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 30 00:36:13.660083 systemd-resolved[260]: Defaulting to hostname 'linux'. Apr 30 00:36:13.664602 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 30 00:36:13.680932 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 30 00:36:13.787652 kernel: SCSI subsystem initialized Apr 30 00:36:13.794666 kernel: Loading iSCSI transport class v2.0-870. Apr 30 00:36:13.804665 kernel: iscsi: registered transport (tcp) Apr 30 00:36:13.823255 kernel: iscsi: registered transport (qla4xxx) Apr 30 00:36:13.823323 kernel: QLogic iSCSI HBA Driver Apr 30 00:36:13.863715 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 30 00:36:13.889811 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 30 00:36:13.923709 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 30 00:36:13.923755 kernel: device-mapper: uevent: version 1.0.3 Apr 30 00:36:13.931189 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Apr 30 00:36:13.983666 kernel: raid6: neonx8 gen() 15741 MB/s Apr 30 00:36:14.002649 kernel: raid6: neonx4 gen() 15654 MB/s Apr 30 00:36:14.022645 kernel: raid6: neonx2 gen() 13259 MB/s Apr 30 00:36:14.046651 kernel: raid6: neonx1 gen() 10492 MB/s Apr 30 00:36:14.066646 kernel: raid6: int64x8 gen() 6949 MB/s Apr 30 00:36:14.087647 kernel: raid6: int64x4 gen() 7344 MB/s Apr 30 00:36:14.107645 kernel: raid6: int64x2 gen() 6131 MB/s Apr 30 00:36:14.130578 kernel: raid6: int64x1 gen() 5059 MB/s Apr 30 00:36:14.130595 kernel: raid6: using algorithm neonx8 gen() 15741 MB/s Apr 30 00:36:14.155597 kernel: raid6: .... xor() 11925 MB/s, rmw enabled Apr 30 00:36:14.155612 kernel: raid6: using neon recovery algorithm Apr 30 00:36:14.163652 kernel: xor: measuring software checksum speed Apr 30 00:36:14.170171 kernel: 8regs : 18444 MB/sec Apr 30 00:36:14.170243 kernel: 32regs : 19627 MB/sec Apr 30 00:36:14.173544 kernel: arm64_neon : 27087 MB/sec Apr 30 00:36:14.177362 kernel: xor: using function: arm64_neon (27087 MB/sec) Apr 30 00:36:14.229661 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 30 00:36:14.241165 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 30 00:36:14.256790 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 00:36:14.279530 systemd-udevd[439]: Using default interface naming scheme 'v255'. Apr 30 00:36:14.285113 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 00:36:14.309803 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 30 00:36:14.327749 dracut-pre-trigger[450]: rd.md=0: removing MD RAID activation Apr 30 00:36:14.357408 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 30 00:36:14.371884 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 30 00:36:14.408570 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 00:36:14.428960 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 30 00:36:14.459556 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 30 00:36:14.474704 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 30 00:36:14.491266 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 00:36:14.506028 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 30 00:36:14.522916 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 30 00:36:14.544798 kernel: hv_vmbus: Vmbus version:5.3 Apr 30 00:36:14.557026 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 30 00:36:14.572182 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 30 00:36:14.582792 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 00:36:14.639947 kernel: hv_vmbus: registering driver hid_hyperv Apr 30 00:36:14.639970 kernel: hv_vmbus: registering driver hyperv_keyboard Apr 30 00:36:14.639990 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input0 Apr 30 00:36:14.640000 kernel: pps_core: LinuxPPS API ver. 1 registered Apr 30 00:36:14.640009 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Apr 30 00:36:14.640141 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input1 Apr 30 00:36:14.640151 kernel: hv_vmbus: registering driver hv_netvsc Apr 30 00:36:14.640160 kernel: hv_vmbus: registering driver hv_storvsc Apr 30 00:36:14.640169 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Apr 30 00:36:14.640178 kernel: scsi host1: storvsc_host_t Apr 30 00:36:14.653517 kernel: scsi host0: storvsc_host_t Apr 30 00:36:14.656257 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 00:36:14.678384 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Apr 30 00:36:14.669843 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 00:36:14.670037 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:14.684881 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 00:36:14.724548 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Apr 30 00:36:14.724768 kernel: PTP clock support registered Apr 30 00:36:14.718442 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 00:36:14.739189 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 00:36:14.739289 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:14.787958 kernel: hv_utils: Registering HyperV Utility Driver Apr 30 00:36:14.787982 kernel: hv_vmbus: registering driver hv_utils Apr 30 00:36:14.787992 kernel: hv_netvsc 002248b7-44a6-0022-48b7-44a6002248b7 eth0: VF slot 1 added Apr 30 00:36:14.788133 kernel: hv_utils: Heartbeat IC version 3.0 Apr 30 00:36:14.788144 kernel: hv_utils: Shutdown IC version 3.2 Apr 30 00:36:14.788153 kernel: hv_utils: TimeSync IC version 4.0 Apr 30 00:36:14.784921 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 00:36:14.612549 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Apr 30 00:36:14.638007 kernel: hv_vmbus: registering driver hv_pci Apr 30 00:36:14.638024 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 30 00:36:14.638035 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Apr 30 00:36:14.640562 kernel: hv_pci 4fff5134-7a3a-45d7-86c2-84969e263b47: PCI VMBus probing: Using version 0x10004 Apr 30 00:36:14.943420 kernel: hv_pci 4fff5134-7a3a-45d7-86c2-84969e263b47: PCI host bridge to bus 7a3a:00 Apr 30 00:36:14.943538 systemd-journald[217]: Time jumped backwards, rotating. Apr 30 00:36:14.943590 kernel: pci_bus 7a3a:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Apr 30 00:36:14.943695 kernel: pci_bus 7a3a:00: No busn resource found for root bus, will use [bus 00-ff] Apr 30 00:36:14.943775 kernel: pci 7a3a:00:02.0: [15b3:1018] type 00 class 0x020000 Apr 30 00:36:14.943885 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Apr 30 00:36:14.943981 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Apr 30 00:36:14.944097 kernel: pci 7a3a:00:02.0: reg 0x10: [mem 0xfc0000000-0xfc00fffff 64bit pref] Apr 30 00:36:14.944204 kernel: sd 0:0:0:0: [sda] Write Protect is off Apr 30 00:36:14.944298 kernel: pci 7a3a:00:02.0: enabling Extended Tags Apr 30 00:36:14.944400 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Apr 30 00:36:14.944485 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Apr 30 00:36:14.944567 kernel: pci 7a3a:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 7a3a:00:02.0 (capable of 126.016 Gb/s with 8.0 GT/s PCIe x16 link) Apr 30 00:36:14.944657 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 30 00:36:14.944666 kernel: pci_bus 7a3a:00: busn_res: [bus 00-ff] end is updated to 00 Apr 30 00:36:14.944744 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Apr 30 00:36:14.944828 kernel: pci 7a3a:00:02.0: BAR 0: assigned [mem 0xfc0000000-0xfc00fffff 64bit pref] Apr 30 00:36:14.585985 systemd-resolved[260]: Clock change detected. Flushing caches. Apr 30 00:36:14.637750 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:14.665579 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 00:36:14.910607 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 00:36:14.987188 kernel: mlx5_core 7a3a:00:02.0: enabling device (0000 -> 0002) Apr 30 00:36:15.722262 kernel: mlx5_core 7a3a:00:02.0: firmware version: 16.31.2424 Apr 30 00:36:15.722441 kernel: hv_netvsc 002248b7-44a6-0022-48b7-44a6002248b7 eth0: VF registering: eth1 Apr 30 00:36:15.722566 kernel: mlx5_core 7a3a:00:02.0 eth1: joined to eth0 Apr 30 00:36:15.722731 kernel: mlx5_core 7a3a:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Apr 30 00:36:15.732099 kernel: mlx5_core 7a3a:00:02.0 enP31290s1: renamed from eth1 Apr 30 00:36:15.966057 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Apr 30 00:36:16.071097 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by (udev-worker) (489) Apr 30 00:36:16.085823 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Apr 30 00:36:16.099333 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Apr 30 00:36:16.120795 kernel: BTRFS: device fsid 7216ceb7-401c-42de-84de-44adb68241e4 devid 1 transid 39 /dev/sda3 scanned by (udev-worker) (494) Apr 30 00:36:16.131040 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Apr 30 00:36:16.147345 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Apr 30 00:36:16.170280 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 30 00:36:16.194141 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 30 00:36:16.202106 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 30 00:36:17.211162 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 30 00:36:17.211217 disk-uuid[602]: The operation has completed successfully. Apr 30 00:36:17.276231 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 30 00:36:17.276352 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 30 00:36:17.305228 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 30 00:36:17.317275 sh[688]: Success Apr 30 00:36:17.347323 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Apr 30 00:36:17.537062 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 30 00:36:17.557233 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 30 00:36:17.563097 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 30 00:36:17.595986 kernel: BTRFS info (device dm-0): first mount of filesystem 7216ceb7-401c-42de-84de-44adb68241e4 Apr 30 00:36:17.596044 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Apr 30 00:36:17.600096 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Apr 30 00:36:17.607657 kernel: BTRFS info (device dm-0): disabling log replay at mount time Apr 30 00:36:17.613113 kernel: BTRFS info (device dm-0): using free space tree Apr 30 00:36:17.896818 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 30 00:36:17.902393 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 30 00:36:17.922331 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 30 00:36:17.934283 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 30 00:36:17.965719 kernel: BTRFS info (device sda6): first mount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:17.965741 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 30 00:36:17.965759 kernel: BTRFS info (device sda6): using free space tree Apr 30 00:36:17.986126 kernel: BTRFS info (device sda6): auto enabling async discard Apr 30 00:36:17.999536 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 30 00:36:18.007451 kernel: BTRFS info (device sda6): last unmount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:18.017248 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 30 00:36:18.031358 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 30 00:36:18.098737 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 30 00:36:18.118336 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 30 00:36:18.145532 systemd-networkd[872]: lo: Link UP Apr 30 00:36:18.145543 systemd-networkd[872]: lo: Gained carrier Apr 30 00:36:18.147262 systemd-networkd[872]: Enumeration completed Apr 30 00:36:18.147366 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 30 00:36:18.155881 systemd-networkd[872]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 00:36:18.155884 systemd-networkd[872]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 00:36:18.156424 systemd[1]: Reached target network.target - Network. Apr 30 00:36:18.217090 kernel: mlx5_core 7a3a:00:02.0 enP31290s1: Link up Apr 30 00:36:18.298089 kernel: hv_netvsc 002248b7-44a6-0022-48b7-44a6002248b7 eth0: Data path switched to VF: enP31290s1 Apr 30 00:36:18.298602 systemd-networkd[872]: enP31290s1: Link UP Apr 30 00:36:18.298691 systemd-networkd[872]: eth0: Link UP Apr 30 00:36:18.298787 systemd-networkd[872]: eth0: Gained carrier Apr 30 00:36:18.298797 systemd-networkd[872]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 00:36:18.322214 systemd-networkd[872]: enP31290s1: Gained carrier Apr 30 00:36:18.331119 systemd-networkd[872]: eth0: DHCPv4 address 10.200.20.12/24, gateway 10.200.20.1 acquired from 168.63.129.16 Apr 30 00:36:19.022508 ignition[788]: Ignition 2.19.0 Apr 30 00:36:19.022519 ignition[788]: Stage: fetch-offline Apr 30 00:36:19.027382 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 30 00:36:19.022560 ignition[788]: no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:19.022568 ignition[788]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:19.022665 ignition[788]: parsed url from cmdline: "" Apr 30 00:36:19.022668 ignition[788]: no config URL provided Apr 30 00:36:19.022672 ignition[788]: reading system config file "/usr/lib/ignition/user.ign" Apr 30 00:36:19.053375 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Apr 30 00:36:19.022678 ignition[788]: no config at "/usr/lib/ignition/user.ign" Apr 30 00:36:19.022683 ignition[788]: failed to fetch config: resource requires networking Apr 30 00:36:19.022869 ignition[788]: Ignition finished successfully Apr 30 00:36:19.073333 ignition[882]: Ignition 2.19.0 Apr 30 00:36:19.073340 ignition[882]: Stage: fetch Apr 30 00:36:19.073543 ignition[882]: no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:19.073554 ignition[882]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:19.073661 ignition[882]: parsed url from cmdline: "" Apr 30 00:36:19.073664 ignition[882]: no config URL provided Apr 30 00:36:19.073668 ignition[882]: reading system config file "/usr/lib/ignition/user.ign" Apr 30 00:36:19.073676 ignition[882]: no config at "/usr/lib/ignition/user.ign" Apr 30 00:36:19.073707 ignition[882]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Apr 30 00:36:19.185421 ignition[882]: GET result: OK Apr 30 00:36:19.185541 ignition[882]: config has been read from IMDS userdata Apr 30 00:36:19.185581 ignition[882]: parsing config with SHA512: c74a6cc6e4182038b30a5bb84c7ceca9605595d73dfdb51b912378251a996c0afd24e625706b4d4b5a3447b7f04c1949bc5afeef0132bd898fd6fb0ae5d80175 Apr 30 00:36:19.189590 unknown[882]: fetched base config from "system" Apr 30 00:36:19.189986 ignition[882]: fetch: fetch complete Apr 30 00:36:19.189597 unknown[882]: fetched base config from "system" Apr 30 00:36:19.189991 ignition[882]: fetch: fetch passed Apr 30 00:36:19.189602 unknown[882]: fetched user config from "azure" Apr 30 00:36:19.190031 ignition[882]: Ignition finished successfully Apr 30 00:36:19.192202 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Apr 30 00:36:19.215373 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 30 00:36:19.242955 ignition[890]: Ignition 2.19.0 Apr 30 00:36:19.242970 ignition[890]: Stage: kargs Apr 30 00:36:19.243185 ignition[890]: no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:19.250324 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 30 00:36:19.243198 ignition[890]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:19.244161 ignition[890]: kargs: kargs passed Apr 30 00:36:19.244210 ignition[890]: Ignition finished successfully Apr 30 00:36:19.277364 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 30 00:36:19.295718 ignition[896]: Ignition 2.19.0 Apr 30 00:36:19.295730 ignition[896]: Stage: disks Apr 30 00:36:19.301127 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 30 00:36:19.295901 ignition[896]: no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:19.308754 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 30 00:36:19.295910 ignition[896]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:19.319219 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 30 00:36:19.296845 ignition[896]: disks: disks passed Apr 30 00:36:19.330355 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 30 00:36:19.296895 ignition[896]: Ignition finished successfully Apr 30 00:36:19.341783 systemd[1]: Reached target sysinit.target - System Initialization. Apr 30 00:36:19.354928 systemd[1]: Reached target basic.target - Basic System. Apr 30 00:36:19.367029 systemd-networkd[872]: eth0: Gained IPv6LL Apr 30 00:36:19.367391 systemd-networkd[872]: enP31290s1: Gained IPv6LL Apr 30 00:36:19.384354 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 30 00:36:19.477835 systemd-fsck[904]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Apr 30 00:36:19.487211 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 30 00:36:19.504354 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 30 00:36:19.567091 kernel: EXT4-fs (sda9): mounted filesystem c13301f3-70ec-4948-963a-f1db0e953273 r/w with ordered data mode. Quota mode: none. Apr 30 00:36:19.567476 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 30 00:36:19.572740 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 30 00:36:19.625188 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 30 00:36:19.632238 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 30 00:36:19.643250 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Apr 30 00:36:19.657563 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 30 00:36:19.657599 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 30 00:36:19.665570 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 30 00:36:19.707262 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (915) Apr 30 00:36:19.707322 kernel: BTRFS info (device sda6): first mount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:19.707846 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 30 00:36:19.729716 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 30 00:36:19.729741 kernel: BTRFS info (device sda6): using free space tree Apr 30 00:36:19.736142 kernel: BTRFS info (device sda6): auto enabling async discard Apr 30 00:36:19.736981 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 30 00:36:20.212952 coreos-metadata[917]: Apr 30 00:36:20.212 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Apr 30 00:36:20.222900 coreos-metadata[917]: Apr 30 00:36:20.222 INFO Fetch successful Apr 30 00:36:20.222900 coreos-metadata[917]: Apr 30 00:36:20.222 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Apr 30 00:36:20.239426 coreos-metadata[917]: Apr 30 00:36:20.234 INFO Fetch successful Apr 30 00:36:20.239426 coreos-metadata[917]: Apr 30 00:36:20.235 INFO wrote hostname ci-4081.3.3-a-c388cdac8a to /sysroot/etc/hostname Apr 30 00:36:20.239809 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 30 00:36:20.547677 initrd-setup-root[945]: cut: /sysroot/etc/passwd: No such file or directory Apr 30 00:36:20.625038 initrd-setup-root[952]: cut: /sysroot/etc/group: No such file or directory Apr 30 00:36:20.647548 initrd-setup-root[959]: cut: /sysroot/etc/shadow: No such file or directory Apr 30 00:36:20.657606 initrd-setup-root[966]: cut: /sysroot/etc/gshadow: No such file or directory Apr 30 00:36:21.426933 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 30 00:36:21.440570 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 30 00:36:21.450328 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 30 00:36:21.472188 kernel: BTRFS info (device sda6): last unmount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:21.466805 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 30 00:36:21.496708 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 30 00:36:21.511096 ignition[1035]: INFO : Ignition 2.19.0 Apr 30 00:36:21.511096 ignition[1035]: INFO : Stage: mount Apr 30 00:36:21.525274 ignition[1035]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:21.525274 ignition[1035]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:21.525274 ignition[1035]: INFO : mount: mount passed Apr 30 00:36:21.525274 ignition[1035]: INFO : Ignition finished successfully Apr 30 00:36:21.516269 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 30 00:36:21.542345 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 30 00:36:21.561285 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 30 00:36:21.601774 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (1045) Apr 30 00:36:21.601847 kernel: BTRFS info (device sda6): first mount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:21.608567 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 30 00:36:21.613350 kernel: BTRFS info (device sda6): using free space tree Apr 30 00:36:21.620096 kernel: BTRFS info (device sda6): auto enabling async discard Apr 30 00:36:21.621224 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 30 00:36:21.651254 ignition[1062]: INFO : Ignition 2.19.0 Apr 30 00:36:21.651254 ignition[1062]: INFO : Stage: files Apr 30 00:36:21.659467 ignition[1062]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:21.659467 ignition[1062]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:21.659467 ignition[1062]: DEBUG : files: compiled without relabeling support, skipping Apr 30 00:36:21.682558 ignition[1062]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 30 00:36:21.682558 ignition[1062]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 30 00:36:21.771328 ignition[1062]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 30 00:36:21.779875 ignition[1062]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 30 00:36:21.779875 ignition[1062]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 30 00:36:21.772467 unknown[1062]: wrote ssh authorized keys file for user: core Apr 30 00:36:21.808837 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 30 00:36:21.819274 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Apr 30 00:36:21.899008 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 30 00:36:22.002707 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 30 00:36:22.002707 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-arm64.raw: attempt #1 Apr 30 00:36:22.429357 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Apr 30 00:36:22.689965 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 00:36:22.689965 ignition[1062]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Apr 30 00:36:22.736102 ignition[1062]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: files passed Apr 30 00:36:22.748075 ignition[1062]: INFO : Ignition finished successfully Apr 30 00:36:22.760109 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 30 00:36:22.792392 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 30 00:36:22.811289 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 30 00:36:22.835869 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 30 00:36:22.835976 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 30 00:36:22.885589 initrd-setup-root-after-ignition[1095]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 30 00:36:22.893700 initrd-setup-root-after-ignition[1091]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 30 00:36:22.893700 initrd-setup-root-after-ignition[1091]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 30 00:36:22.885883 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 30 00:36:22.900963 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 30 00:36:22.941424 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 30 00:36:22.976311 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 30 00:36:22.976432 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 30 00:36:22.989183 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 30 00:36:23.000858 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 30 00:36:23.011516 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 30 00:36:23.029362 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 30 00:36:23.051582 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 30 00:36:23.068312 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 30 00:36:23.086593 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 30 00:36:23.093686 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 00:36:23.107426 systemd[1]: Stopped target timers.target - Timer Units. Apr 30 00:36:23.118935 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 30 00:36:23.119140 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 30 00:36:23.135612 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 30 00:36:23.147465 systemd[1]: Stopped target basic.target - Basic System. Apr 30 00:36:23.160313 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 30 00:36:23.171642 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 30 00:36:23.188251 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 30 00:36:23.200533 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 30 00:36:23.211549 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 30 00:36:23.223492 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 30 00:36:23.235683 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 30 00:36:23.246442 systemd[1]: Stopped target swap.target - Swaps. Apr 30 00:36:23.256283 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 30 00:36:23.256466 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 30 00:36:23.274955 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 30 00:36:23.286198 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 00:36:23.298150 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 30 00:36:23.304023 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 00:36:23.310802 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 30 00:36:23.310998 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 30 00:36:23.328635 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 30 00:36:23.328826 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 30 00:36:23.340119 systemd[1]: ignition-files.service: Deactivated successfully. Apr 30 00:36:23.340276 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 30 00:36:23.350327 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Apr 30 00:36:23.350487 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 30 00:36:23.381238 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 30 00:36:23.409184 ignition[1115]: INFO : Ignition 2.19.0 Apr 30 00:36:23.409184 ignition[1115]: INFO : Stage: umount Apr 30 00:36:23.409184 ignition[1115]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:23.409184 ignition[1115]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:23.409184 ignition[1115]: INFO : umount: umount passed Apr 30 00:36:23.409184 ignition[1115]: INFO : Ignition finished successfully Apr 30 00:36:23.390010 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 30 00:36:23.390320 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 00:36:23.429236 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 30 00:36:23.438884 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 30 00:36:23.439461 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 00:36:23.454305 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 30 00:36:23.454483 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 30 00:36:23.473867 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 30 00:36:23.473963 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 30 00:36:23.486327 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 30 00:36:23.486430 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 30 00:36:23.495305 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 30 00:36:23.495362 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 30 00:36:23.513267 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 30 00:36:23.513340 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 30 00:36:23.524492 systemd[1]: ignition-fetch.service: Deactivated successfully. Apr 30 00:36:23.524560 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Apr 30 00:36:23.530887 systemd[1]: Stopped target network.target - Network. Apr 30 00:36:23.540840 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 30 00:36:23.540919 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 30 00:36:23.551778 systemd[1]: Stopped target paths.target - Path Units. Apr 30 00:36:23.562107 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 30 00:36:23.566949 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 00:36:23.574113 systemd[1]: Stopped target slices.target - Slice Units. Apr 30 00:36:23.585641 systemd[1]: Stopped target sockets.target - Socket Units. Apr 30 00:36:23.595632 systemd[1]: iscsid.socket: Deactivated successfully. Apr 30 00:36:23.595696 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 30 00:36:23.605309 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 30 00:36:23.605354 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 30 00:36:23.617301 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 30 00:36:23.617370 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 30 00:36:23.627836 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 30 00:36:23.627899 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 30 00:36:23.638157 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 30 00:36:23.851159 kernel: hv_netvsc 002248b7-44a6-0022-48b7-44a6002248b7 eth0: Data path switched from VF: enP31290s1 Apr 30 00:36:23.648785 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 30 00:36:23.660255 systemd-networkd[872]: eth0: DHCPv6 lease lost Apr 30 00:36:23.662486 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 30 00:36:23.663844 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 30 00:36:23.664269 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 30 00:36:23.674997 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 30 00:36:23.675037 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 30 00:36:23.699205 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 30 00:36:23.710839 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 30 00:36:23.710923 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 30 00:36:23.723583 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 00:36:23.739859 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 30 00:36:23.739986 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 30 00:36:23.757229 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 30 00:36:23.757377 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 30 00:36:23.768427 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 30 00:36:23.768497 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 30 00:36:23.774719 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 30 00:36:23.774773 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 00:36:23.788529 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 30 00:36:23.790115 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 00:36:23.801557 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 30 00:36:23.801640 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 30 00:36:23.812876 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 30 00:36:23.812925 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 00:36:23.824459 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 30 00:36:23.824523 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 30 00:36:23.846248 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 30 00:36:23.846363 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 30 00:36:23.864823 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 30 00:36:23.864892 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 00:36:23.890377 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 30 00:36:23.904256 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 30 00:36:23.904349 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 00:36:23.928892 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 00:36:23.928962 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:23.942128 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 30 00:36:23.942248 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 30 00:36:24.002227 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 30 00:36:24.002355 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 30 00:36:24.119914 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 30 00:36:24.120116 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 30 00:36:24.126528 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 30 00:36:24.138040 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 30 00:36:24.138125 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 30 00:36:24.166397 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 30 00:36:24.185638 systemd[1]: Switching root. Apr 30 00:36:24.290659 systemd-journald[217]: Journal stopped Apr 30 00:36:13.305801 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Apr 30 00:36:13.305824 kernel: Linux version 6.6.88-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Tue Apr 29 23:08:45 -00 2025 Apr 30 00:36:13.305832 kernel: KASLR enabled Apr 30 00:36:13.305838 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Apr 30 00:36:13.305845 kernel: printk: bootconsole [pl11] enabled Apr 30 00:36:13.305850 kernel: efi: EFI v2.7 by EDK II Apr 30 00:36:13.305858 kernel: efi: ACPI 2.0=0x3fd5f018 SMBIOS=0x3e580000 SMBIOS 3.0=0x3e560000 MEMATTR=0x3ead8b98 RNG=0x3fd5f998 MEMRESERVE=0x3e44ee18 Apr 30 00:36:13.305864 kernel: random: crng init done Apr 30 00:36:13.305870 kernel: ACPI: Early table checksum verification disabled Apr 30 00:36:13.305875 kernel: ACPI: RSDP 0x000000003FD5F018 000024 (v02 VRTUAL) Apr 30 00:36:13.305881 kernel: ACPI: XSDT 0x000000003FD5FF18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305887 kernel: ACPI: FACP 0x000000003FD5FC18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305894 kernel: ACPI: DSDT 0x000000003FD41018 01DFCD (v02 MSFTVM DSDT01 00000001 INTL 20230628) Apr 30 00:36:13.305900 kernel: ACPI: DBG2 0x000000003FD5FB18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305907 kernel: ACPI: GTDT 0x000000003FD5FD98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305914 kernel: ACPI: OEM0 0x000000003FD5F098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305920 kernel: ACPI: SPCR 0x000000003FD5FA98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305928 kernel: ACPI: APIC 0x000000003FD5F818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305934 kernel: ACPI: SRAT 0x000000003FD5F198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305940 kernel: ACPI: PPTT 0x000000003FD5F418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Apr 30 00:36:13.305947 kernel: ACPI: BGRT 0x000000003FD5FE98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 30 00:36:13.305953 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Apr 30 00:36:13.305959 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] Apr 30 00:36:13.305965 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] Apr 30 00:36:13.305972 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] Apr 30 00:36:13.305978 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] Apr 30 00:36:13.305984 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] Apr 30 00:36:13.305990 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] Apr 30 00:36:13.305998 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] Apr 30 00:36:13.306004 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] Apr 30 00:36:13.306010 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] Apr 30 00:36:13.306016 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] Apr 30 00:36:13.306023 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] Apr 30 00:36:13.306029 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] Apr 30 00:36:13.306035 kernel: NUMA: NODE_DATA [mem 0x1bf7ee800-0x1bf7f3fff] Apr 30 00:36:13.306041 kernel: Zone ranges: Apr 30 00:36:13.306048 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Apr 30 00:36:13.306054 kernel: DMA32 empty Apr 30 00:36:13.306060 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Apr 30 00:36:13.306066 kernel: Movable zone start for each node Apr 30 00:36:13.306077 kernel: Early memory node ranges Apr 30 00:36:13.306083 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Apr 30 00:36:13.306090 kernel: node 0: [mem 0x0000000000824000-0x000000003e54ffff] Apr 30 00:36:13.306096 kernel: node 0: [mem 0x000000003e550000-0x000000003e87ffff] Apr 30 00:36:13.306103 kernel: node 0: [mem 0x000000003e880000-0x000000003fc7ffff] Apr 30 00:36:13.306110 kernel: node 0: [mem 0x000000003fc80000-0x000000003fcfffff] Apr 30 00:36:13.306117 kernel: node 0: [mem 0x000000003fd00000-0x000000003fffffff] Apr 30 00:36:13.306124 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Apr 30 00:36:13.306130 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Apr 30 00:36:13.306137 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Apr 30 00:36:13.306144 kernel: psci: probing for conduit method from ACPI. Apr 30 00:36:13.306150 kernel: psci: PSCIv1.1 detected in firmware. Apr 30 00:36:13.306157 kernel: psci: Using standard PSCI v0.2 function IDs Apr 30 00:36:13.306163 kernel: psci: MIGRATE_INFO_TYPE not supported. Apr 30 00:36:13.306170 kernel: psci: SMC Calling Convention v1.4 Apr 30 00:36:13.306176 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Apr 30 00:36:13.306183 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Apr 30 00:36:13.306191 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Apr 30 00:36:13.306198 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Apr 30 00:36:13.306205 kernel: pcpu-alloc: [0] 0 [0] 1 Apr 30 00:36:13.306211 kernel: Detected PIPT I-cache on CPU0 Apr 30 00:36:13.306218 kernel: CPU features: detected: GIC system register CPU interface Apr 30 00:36:13.306225 kernel: CPU features: detected: Hardware dirty bit management Apr 30 00:36:13.306231 kernel: CPU features: detected: Spectre-BHB Apr 30 00:36:13.306238 kernel: CPU features: kernel page table isolation forced ON by KASLR Apr 30 00:36:13.306245 kernel: CPU features: detected: Kernel page table isolation (KPTI) Apr 30 00:36:13.306251 kernel: CPU features: detected: ARM erratum 1418040 Apr 30 00:36:13.306258 kernel: CPU features: detected: ARM erratum 1542419 (kernel portion) Apr 30 00:36:13.306266 kernel: CPU features: detected: SSBS not fully self-synchronizing Apr 30 00:36:13.306272 kernel: alternatives: applying boot alternatives Apr 30 00:36:13.306280 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=2f2ec97241771b99b21726307071be4f8c5924f9157dc58cd38c4fcfbe71412a Apr 30 00:36:13.306288 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Apr 30 00:36:13.306294 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 30 00:36:13.306301 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 30 00:36:13.306307 kernel: Fallback order for Node 0: 0 Apr 30 00:36:13.306314 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1032156 Apr 30 00:36:13.306321 kernel: Policy zone: Normal Apr 30 00:36:13.306327 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 30 00:36:13.306334 kernel: software IO TLB: area num 2. Apr 30 00:36:13.306342 kernel: software IO TLB: mapped [mem 0x000000003a44e000-0x000000003e44e000] (64MB) Apr 30 00:36:13.306349 kernel: Memory: 3982688K/4194160K available (10240K kernel code, 2186K rwdata, 8104K rodata, 39424K init, 897K bss, 211472K reserved, 0K cma-reserved) Apr 30 00:36:13.306356 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Apr 30 00:36:13.306362 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 30 00:36:13.306370 kernel: rcu: RCU event tracing is enabled. Apr 30 00:36:13.306377 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Apr 30 00:36:13.306384 kernel: Trampoline variant of Tasks RCU enabled. Apr 30 00:36:13.306390 kernel: Tracing variant of Tasks RCU enabled. Apr 30 00:36:13.306397 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 30 00:36:13.306404 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Apr 30 00:36:13.306411 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Apr 30 00:36:13.306418 kernel: GICv3: 960 SPIs implemented Apr 30 00:36:13.306425 kernel: GICv3: 0 Extended SPIs implemented Apr 30 00:36:13.306432 kernel: Root IRQ handler: gic_handle_irq Apr 30 00:36:13.306438 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Apr 30 00:36:13.306445 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Apr 30 00:36:13.306452 kernel: ITS: No ITS available, not enabling LPIs Apr 30 00:36:13.306458 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 30 00:36:13.306465 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 00:36:13.306472 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Apr 30 00:36:13.306478 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Apr 30 00:36:13.306485 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Apr 30 00:36:13.306493 kernel: Console: colour dummy device 80x25 Apr 30 00:36:13.306501 kernel: printk: console [tty1] enabled Apr 30 00:36:13.306507 kernel: ACPI: Core revision 20230628 Apr 30 00:36:13.306514 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Apr 30 00:36:13.306522 kernel: pid_max: default: 32768 minimum: 301 Apr 30 00:36:13.306529 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Apr 30 00:36:13.306535 kernel: landlock: Up and running. Apr 30 00:36:13.306542 kernel: SELinux: Initializing. Apr 30 00:36:13.306549 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 30 00:36:13.306556 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 30 00:36:13.306564 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 30 00:36:13.306571 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 30 00:36:13.306578 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3a8030, hints 0xe, misc 0x31e1 Apr 30 00:36:13.306585 kernel: Hyper-V: Host Build 10.0.22477.1619-1-0 Apr 30 00:36:13.306591 kernel: Hyper-V: enabling crash_kexec_post_notifiers Apr 30 00:36:13.306598 kernel: rcu: Hierarchical SRCU implementation. Apr 30 00:36:13.306605 kernel: rcu: Max phase no-delay instances is 400. Apr 30 00:36:13.306618 kernel: Remapping and enabling EFI services. Apr 30 00:36:13.306626 kernel: smp: Bringing up secondary CPUs ... Apr 30 00:36:13.306647 kernel: Detected PIPT I-cache on CPU1 Apr 30 00:36:13.306655 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Apr 30 00:36:13.306664 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 00:36:13.306671 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Apr 30 00:36:13.306678 kernel: smp: Brought up 1 node, 2 CPUs Apr 30 00:36:13.306686 kernel: SMP: Total of 2 processors activated. Apr 30 00:36:13.306693 kernel: CPU features: detected: 32-bit EL0 Support Apr 30 00:36:13.306701 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Apr 30 00:36:13.306709 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Apr 30 00:36:13.306716 kernel: CPU features: detected: CRC32 instructions Apr 30 00:36:13.306723 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Apr 30 00:36:13.306730 kernel: CPU features: detected: LSE atomic instructions Apr 30 00:36:13.306738 kernel: CPU features: detected: Privileged Access Never Apr 30 00:36:13.306745 kernel: CPU: All CPU(s) started at EL1 Apr 30 00:36:13.306752 kernel: alternatives: applying system-wide alternatives Apr 30 00:36:13.306759 kernel: devtmpfs: initialized Apr 30 00:36:13.306768 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 30 00:36:13.306775 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Apr 30 00:36:13.306783 kernel: pinctrl core: initialized pinctrl subsystem Apr 30 00:36:13.306790 kernel: SMBIOS 3.1.0 present. Apr 30 00:36:13.306797 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 09/28/2024 Apr 30 00:36:13.306804 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 30 00:36:13.306812 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Apr 30 00:36:13.306819 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Apr 30 00:36:13.306826 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Apr 30 00:36:13.306835 kernel: audit: initializing netlink subsys (disabled) Apr 30 00:36:13.306842 kernel: audit: type=2000 audit(0.047:1): state=initialized audit_enabled=0 res=1 Apr 30 00:36:13.306850 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 30 00:36:13.306857 kernel: cpuidle: using governor menu Apr 30 00:36:13.306864 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Apr 30 00:36:13.306871 kernel: ASID allocator initialised with 32768 entries Apr 30 00:36:13.306879 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 30 00:36:13.306886 kernel: Serial: AMBA PL011 UART driver Apr 30 00:36:13.306893 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Apr 30 00:36:13.306902 kernel: Modules: 0 pages in range for non-PLT usage Apr 30 00:36:13.306909 kernel: Modules: 509024 pages in range for PLT usage Apr 30 00:36:13.306917 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 30 00:36:13.306924 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Apr 30 00:36:13.306931 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Apr 30 00:36:13.306938 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Apr 30 00:36:13.306945 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 30 00:36:13.306953 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Apr 30 00:36:13.306960 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Apr 30 00:36:13.306968 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Apr 30 00:36:13.306976 kernel: ACPI: Added _OSI(Module Device) Apr 30 00:36:13.306983 kernel: ACPI: Added _OSI(Processor Device) Apr 30 00:36:13.306990 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Apr 30 00:36:13.306997 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 30 00:36:13.307004 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 30 00:36:13.307011 kernel: ACPI: Interpreter enabled Apr 30 00:36:13.307018 kernel: ACPI: Using GIC for interrupt routing Apr 30 00:36:13.307026 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Apr 30 00:36:13.307034 kernel: printk: console [ttyAMA0] enabled Apr 30 00:36:13.307041 kernel: printk: bootconsole [pl11] disabled Apr 30 00:36:13.307049 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Apr 30 00:36:13.307056 kernel: iommu: Default domain type: Translated Apr 30 00:36:13.307063 kernel: iommu: DMA domain TLB invalidation policy: strict mode Apr 30 00:36:13.307070 kernel: efivars: Registered efivars operations Apr 30 00:36:13.307078 kernel: vgaarb: loaded Apr 30 00:36:13.307085 kernel: clocksource: Switched to clocksource arch_sys_counter Apr 30 00:36:13.307092 kernel: VFS: Disk quotas dquot_6.6.0 Apr 30 00:36:13.307101 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 30 00:36:13.307108 kernel: pnp: PnP ACPI init Apr 30 00:36:13.307115 kernel: pnp: PnP ACPI: found 0 devices Apr 30 00:36:13.307122 kernel: NET: Registered PF_INET protocol family Apr 30 00:36:13.307129 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 30 00:36:13.307137 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 30 00:36:13.307144 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 30 00:36:13.307151 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 30 00:36:13.307159 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Apr 30 00:36:13.307167 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 30 00:36:13.307175 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 30 00:36:13.307182 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 30 00:36:13.307189 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 30 00:36:13.307196 kernel: PCI: CLS 0 bytes, default 64 Apr 30 00:36:13.307204 kernel: kvm [1]: HYP mode not available Apr 30 00:36:13.307211 kernel: Initialise system trusted keyrings Apr 30 00:36:13.307218 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 30 00:36:13.307225 kernel: Key type asymmetric registered Apr 30 00:36:13.307233 kernel: Asymmetric key parser 'x509' registered Apr 30 00:36:13.307241 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Apr 30 00:36:13.307248 kernel: io scheduler mq-deadline registered Apr 30 00:36:13.307255 kernel: io scheduler kyber registered Apr 30 00:36:13.307262 kernel: io scheduler bfq registered Apr 30 00:36:13.307270 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 30 00:36:13.307277 kernel: thunder_xcv, ver 1.0 Apr 30 00:36:13.307284 kernel: thunder_bgx, ver 1.0 Apr 30 00:36:13.307291 kernel: nicpf, ver 1.0 Apr 30 00:36:13.307298 kernel: nicvf, ver 1.0 Apr 30 00:36:13.307449 kernel: rtc-efi rtc-efi.0: registered as rtc0 Apr 30 00:36:13.307525 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-04-30T00:36:12 UTC (1745973372) Apr 30 00:36:13.307535 kernel: efifb: probing for efifb Apr 30 00:36:13.307543 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Apr 30 00:36:13.307550 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Apr 30 00:36:13.307557 kernel: efifb: scrolling: redraw Apr 30 00:36:13.307564 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Apr 30 00:36:13.307574 kernel: Console: switching to colour frame buffer device 128x48 Apr 30 00:36:13.307582 kernel: fb0: EFI VGA frame buffer device Apr 30 00:36:13.307589 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Apr 30 00:36:13.307596 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 30 00:36:13.307603 kernel: No ACPI PMU IRQ for CPU0 Apr 30 00:36:13.307611 kernel: No ACPI PMU IRQ for CPU1 Apr 30 00:36:13.307618 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 1 counters available Apr 30 00:36:13.307625 kernel: watchdog: Delayed init of the lockup detector failed: -19 Apr 30 00:36:13.307644 kernel: watchdog: Hard watchdog permanently disabled Apr 30 00:36:13.307653 kernel: NET: Registered PF_INET6 protocol family Apr 30 00:36:13.307660 kernel: Segment Routing with IPv6 Apr 30 00:36:13.307667 kernel: In-situ OAM (IOAM) with IPv6 Apr 30 00:36:13.307675 kernel: NET: Registered PF_PACKET protocol family Apr 30 00:36:13.307682 kernel: Key type dns_resolver registered Apr 30 00:36:13.307689 kernel: registered taskstats version 1 Apr 30 00:36:13.307697 kernel: Loading compiled-in X.509 certificates Apr 30 00:36:13.307704 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.88-flatcar: e2b28159d3a83b6f5d5db45519e470b1b834e378' Apr 30 00:36:13.307711 kernel: Key type .fscrypt registered Apr 30 00:36:13.307720 kernel: Key type fscrypt-provisioning registered Apr 30 00:36:13.307727 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 30 00:36:13.307735 kernel: ima: Allocated hash algorithm: sha1 Apr 30 00:36:13.307742 kernel: ima: No architecture policies found Apr 30 00:36:13.307749 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Apr 30 00:36:13.307756 kernel: clk: Disabling unused clocks Apr 30 00:36:13.307763 kernel: Freeing unused kernel memory: 39424K Apr 30 00:36:13.307771 kernel: Run /init as init process Apr 30 00:36:13.307778 kernel: with arguments: Apr 30 00:36:13.307786 kernel: /init Apr 30 00:36:13.307794 kernel: with environment: Apr 30 00:36:13.307801 kernel: HOME=/ Apr 30 00:36:13.307808 kernel: TERM=linux Apr 30 00:36:13.307814 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Apr 30 00:36:13.307824 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 30 00:36:13.307834 systemd[1]: Detected virtualization microsoft. Apr 30 00:36:13.307842 systemd[1]: Detected architecture arm64. Apr 30 00:36:13.307851 systemd[1]: Running in initrd. Apr 30 00:36:13.307859 systemd[1]: No hostname configured, using default hostname. Apr 30 00:36:13.307866 systemd[1]: Hostname set to . Apr 30 00:36:13.307874 systemd[1]: Initializing machine ID from random generator. Apr 30 00:36:13.307882 systemd[1]: Queued start job for default target initrd.target. Apr 30 00:36:13.307890 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 00:36:13.307898 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 00:36:13.307906 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 30 00:36:13.307916 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 30 00:36:13.307924 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 30 00:36:13.307932 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 30 00:36:13.307941 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 30 00:36:13.307950 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 30 00:36:13.307957 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 00:36:13.307965 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 30 00:36:13.307974 systemd[1]: Reached target paths.target - Path Units. Apr 30 00:36:13.307982 systemd[1]: Reached target slices.target - Slice Units. Apr 30 00:36:13.307990 systemd[1]: Reached target swap.target - Swaps. Apr 30 00:36:13.307998 systemd[1]: Reached target timers.target - Timer Units. Apr 30 00:36:13.308006 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 30 00:36:13.308014 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 30 00:36:13.308021 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 30 00:36:13.308029 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 30 00:36:13.308039 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 30 00:36:13.308047 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 30 00:36:13.308055 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 00:36:13.308063 systemd[1]: Reached target sockets.target - Socket Units. Apr 30 00:36:13.308071 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 30 00:36:13.308078 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 30 00:36:13.308086 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 30 00:36:13.308094 systemd[1]: Starting systemd-fsck-usr.service... Apr 30 00:36:13.308102 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 30 00:36:13.308112 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 30 00:36:13.308138 systemd-journald[217]: Collecting audit messages is disabled. Apr 30 00:36:13.308158 systemd-journald[217]: Journal started Apr 30 00:36:13.308178 systemd-journald[217]: Runtime Journal (/run/log/journal/0998849b57cc425bbb6a442302b8b3b6) is 8.0M, max 78.5M, 70.5M free. Apr 30 00:36:13.319722 systemd-modules-load[218]: Inserted module 'overlay' Apr 30 00:36:13.329969 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 00:36:13.345655 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 30 00:36:13.345703 systemd[1]: Started systemd-journald.service - Journal Service. Apr 30 00:36:13.354671 kernel: Bridge firewalling registered Apr 30 00:36:13.354816 systemd-modules-load[218]: Inserted module 'br_netfilter' Apr 30 00:36:13.360870 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 30 00:36:13.382788 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 00:36:13.389740 systemd[1]: Finished systemd-fsck-usr.service. Apr 30 00:36:13.400084 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 30 00:36:13.409854 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:13.430931 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 00:36:13.439868 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 30 00:36:13.455849 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 30 00:36:13.483866 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 30 00:36:13.493105 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 00:36:13.507698 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 30 00:36:13.525815 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 30 00:36:13.550397 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 00:36:13.566132 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 30 00:36:13.578002 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 30 00:36:13.589221 dracut-cmdline[251]: dracut-dracut-053 Apr 30 00:36:13.593577 dracut-cmdline[251]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=2f2ec97241771b99b21726307071be4f8c5924f9157dc58cd38c4fcfbe71412a Apr 30 00:36:13.624091 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 30 00:36:13.645218 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 00:36:13.657584 systemd-resolved[260]: Positive Trust Anchors: Apr 30 00:36:13.657593 systemd-resolved[260]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 30 00:36:13.657624 systemd-resolved[260]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 30 00:36:13.660083 systemd-resolved[260]: Defaulting to hostname 'linux'. Apr 30 00:36:13.664602 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 30 00:36:13.680932 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 30 00:36:13.787652 kernel: SCSI subsystem initialized Apr 30 00:36:13.794666 kernel: Loading iSCSI transport class v2.0-870. Apr 30 00:36:13.804665 kernel: iscsi: registered transport (tcp) Apr 30 00:36:13.823255 kernel: iscsi: registered transport (qla4xxx) Apr 30 00:36:13.823323 kernel: QLogic iSCSI HBA Driver Apr 30 00:36:13.863715 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 30 00:36:13.889811 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 30 00:36:13.923709 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 30 00:36:13.923755 kernel: device-mapper: uevent: version 1.0.3 Apr 30 00:36:13.931189 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Apr 30 00:36:13.983666 kernel: raid6: neonx8 gen() 15741 MB/s Apr 30 00:36:14.002649 kernel: raid6: neonx4 gen() 15654 MB/s Apr 30 00:36:14.022645 kernel: raid6: neonx2 gen() 13259 MB/s Apr 30 00:36:14.046651 kernel: raid6: neonx1 gen() 10492 MB/s Apr 30 00:36:14.066646 kernel: raid6: int64x8 gen() 6949 MB/s Apr 30 00:36:14.087647 kernel: raid6: int64x4 gen() 7344 MB/s Apr 30 00:36:14.107645 kernel: raid6: int64x2 gen() 6131 MB/s Apr 30 00:36:14.130578 kernel: raid6: int64x1 gen() 5059 MB/s Apr 30 00:36:14.130595 kernel: raid6: using algorithm neonx8 gen() 15741 MB/s Apr 30 00:36:14.155597 kernel: raid6: .... xor() 11925 MB/s, rmw enabled Apr 30 00:36:14.155612 kernel: raid6: using neon recovery algorithm Apr 30 00:36:14.163652 kernel: xor: measuring software checksum speed Apr 30 00:36:14.170171 kernel: 8regs : 18444 MB/sec Apr 30 00:36:14.170243 kernel: 32regs : 19627 MB/sec Apr 30 00:36:14.173544 kernel: arm64_neon : 27087 MB/sec Apr 30 00:36:14.177362 kernel: xor: using function: arm64_neon (27087 MB/sec) Apr 30 00:36:14.229661 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 30 00:36:14.241165 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 30 00:36:14.256790 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 00:36:14.279530 systemd-udevd[439]: Using default interface naming scheme 'v255'. Apr 30 00:36:14.285113 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 00:36:14.309803 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 30 00:36:14.327749 dracut-pre-trigger[450]: rd.md=0: removing MD RAID activation Apr 30 00:36:14.357408 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 30 00:36:14.371884 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 30 00:36:14.408570 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 00:36:14.428960 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 30 00:36:14.459556 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 30 00:36:14.474704 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 30 00:36:14.491266 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 00:36:14.506028 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 30 00:36:14.522916 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 30 00:36:14.544798 kernel: hv_vmbus: Vmbus version:5.3 Apr 30 00:36:14.557026 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 30 00:36:14.572182 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 30 00:36:14.582792 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 00:36:14.639947 kernel: hv_vmbus: registering driver hid_hyperv Apr 30 00:36:14.639970 kernel: hv_vmbus: registering driver hyperv_keyboard Apr 30 00:36:14.639990 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input0 Apr 30 00:36:14.640000 kernel: pps_core: LinuxPPS API ver. 1 registered Apr 30 00:36:14.640009 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Apr 30 00:36:14.640141 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input1 Apr 30 00:36:14.640151 kernel: hv_vmbus: registering driver hv_netvsc Apr 30 00:36:14.640160 kernel: hv_vmbus: registering driver hv_storvsc Apr 30 00:36:14.640169 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Apr 30 00:36:14.640178 kernel: scsi host1: storvsc_host_t Apr 30 00:36:14.653517 kernel: scsi host0: storvsc_host_t Apr 30 00:36:14.656257 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 00:36:14.678384 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Apr 30 00:36:14.669843 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 00:36:14.670037 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:14.684881 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 00:36:14.724548 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Apr 30 00:36:14.724768 kernel: PTP clock support registered Apr 30 00:36:14.718442 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 00:36:14.739189 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 00:36:14.739289 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:14.787958 kernel: hv_utils: Registering HyperV Utility Driver Apr 30 00:36:14.787982 kernel: hv_vmbus: registering driver hv_utils Apr 30 00:36:14.787992 kernel: hv_netvsc 002248b7-44a6-0022-48b7-44a6002248b7 eth0: VF slot 1 added Apr 30 00:36:14.788133 kernel: hv_utils: Heartbeat IC version 3.0 Apr 30 00:36:14.788144 kernel: hv_utils: Shutdown IC version 3.2 Apr 30 00:36:14.788153 kernel: hv_utils: TimeSync IC version 4.0 Apr 30 00:36:14.784921 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 00:36:14.612549 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Apr 30 00:36:14.638007 kernel: hv_vmbus: registering driver hv_pci Apr 30 00:36:14.638024 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 30 00:36:14.638035 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Apr 30 00:36:14.640562 kernel: hv_pci 4fff5134-7a3a-45d7-86c2-84969e263b47: PCI VMBus probing: Using version 0x10004 Apr 30 00:36:14.943420 kernel: hv_pci 4fff5134-7a3a-45d7-86c2-84969e263b47: PCI host bridge to bus 7a3a:00 Apr 30 00:36:14.943538 systemd-journald[217]: Time jumped backwards, rotating. Apr 30 00:36:14.943590 kernel: pci_bus 7a3a:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Apr 30 00:36:14.943695 kernel: pci_bus 7a3a:00: No busn resource found for root bus, will use [bus 00-ff] Apr 30 00:36:14.943775 kernel: pci 7a3a:00:02.0: [15b3:1018] type 00 class 0x020000 Apr 30 00:36:14.943885 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Apr 30 00:36:14.943981 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Apr 30 00:36:14.944097 kernel: pci 7a3a:00:02.0: reg 0x10: [mem 0xfc0000000-0xfc00fffff 64bit pref] Apr 30 00:36:14.944204 kernel: sd 0:0:0:0: [sda] Write Protect is off Apr 30 00:36:14.944298 kernel: pci 7a3a:00:02.0: enabling Extended Tags Apr 30 00:36:14.944400 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Apr 30 00:36:14.944485 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Apr 30 00:36:14.944567 kernel: pci 7a3a:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 7a3a:00:02.0 (capable of 126.016 Gb/s with 8.0 GT/s PCIe x16 link) Apr 30 00:36:14.944657 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 30 00:36:14.944666 kernel: pci_bus 7a3a:00: busn_res: [bus 00-ff] end is updated to 00 Apr 30 00:36:14.944744 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Apr 30 00:36:14.944828 kernel: pci 7a3a:00:02.0: BAR 0: assigned [mem 0xfc0000000-0xfc00fffff 64bit pref] Apr 30 00:36:14.585985 systemd-resolved[260]: Clock change detected. Flushing caches. Apr 30 00:36:14.637750 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:14.665579 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 00:36:14.910607 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 00:36:14.987188 kernel: mlx5_core 7a3a:00:02.0: enabling device (0000 -> 0002) Apr 30 00:36:15.722262 kernel: mlx5_core 7a3a:00:02.0: firmware version: 16.31.2424 Apr 30 00:36:15.722441 kernel: hv_netvsc 002248b7-44a6-0022-48b7-44a6002248b7 eth0: VF registering: eth1 Apr 30 00:36:15.722566 kernel: mlx5_core 7a3a:00:02.0 eth1: joined to eth0 Apr 30 00:36:15.722731 kernel: mlx5_core 7a3a:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Apr 30 00:36:15.732099 kernel: mlx5_core 7a3a:00:02.0 enP31290s1: renamed from eth1 Apr 30 00:36:15.966057 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Apr 30 00:36:16.071097 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by (udev-worker) (489) Apr 30 00:36:16.085823 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Apr 30 00:36:16.099333 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Apr 30 00:36:16.120795 kernel: BTRFS: device fsid 7216ceb7-401c-42de-84de-44adb68241e4 devid 1 transid 39 /dev/sda3 scanned by (udev-worker) (494) Apr 30 00:36:16.131040 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Apr 30 00:36:16.147345 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Apr 30 00:36:16.170280 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 30 00:36:16.194141 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 30 00:36:16.202106 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 30 00:36:17.211162 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 30 00:36:17.211217 disk-uuid[602]: The operation has completed successfully. Apr 30 00:36:17.276231 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 30 00:36:17.276352 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 30 00:36:17.305228 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 30 00:36:17.317275 sh[688]: Success Apr 30 00:36:17.347323 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Apr 30 00:36:17.537062 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 30 00:36:17.557233 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 30 00:36:17.563097 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 30 00:36:17.595986 kernel: BTRFS info (device dm-0): first mount of filesystem 7216ceb7-401c-42de-84de-44adb68241e4 Apr 30 00:36:17.596044 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Apr 30 00:36:17.600096 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Apr 30 00:36:17.607657 kernel: BTRFS info (device dm-0): disabling log replay at mount time Apr 30 00:36:17.613113 kernel: BTRFS info (device dm-0): using free space tree Apr 30 00:36:17.896818 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 30 00:36:17.902393 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 30 00:36:17.922331 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 30 00:36:17.934283 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 30 00:36:17.965719 kernel: BTRFS info (device sda6): first mount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:17.965741 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 30 00:36:17.965759 kernel: BTRFS info (device sda6): using free space tree Apr 30 00:36:17.986126 kernel: BTRFS info (device sda6): auto enabling async discard Apr 30 00:36:17.999536 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 30 00:36:18.007451 kernel: BTRFS info (device sda6): last unmount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:18.017248 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 30 00:36:18.031358 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 30 00:36:18.098737 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 30 00:36:18.118336 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 30 00:36:18.145532 systemd-networkd[872]: lo: Link UP Apr 30 00:36:18.145543 systemd-networkd[872]: lo: Gained carrier Apr 30 00:36:18.147262 systemd-networkd[872]: Enumeration completed Apr 30 00:36:18.147366 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 30 00:36:18.155881 systemd-networkd[872]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 00:36:18.155884 systemd-networkd[872]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 00:36:18.156424 systemd[1]: Reached target network.target - Network. Apr 30 00:36:18.217090 kernel: mlx5_core 7a3a:00:02.0 enP31290s1: Link up Apr 30 00:36:18.298089 kernel: hv_netvsc 002248b7-44a6-0022-48b7-44a6002248b7 eth0: Data path switched to VF: enP31290s1 Apr 30 00:36:18.298602 systemd-networkd[872]: enP31290s1: Link UP Apr 30 00:36:18.298691 systemd-networkd[872]: eth0: Link UP Apr 30 00:36:18.298787 systemd-networkd[872]: eth0: Gained carrier Apr 30 00:36:18.298797 systemd-networkd[872]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 00:36:18.322214 systemd-networkd[872]: enP31290s1: Gained carrier Apr 30 00:36:18.331119 systemd-networkd[872]: eth0: DHCPv4 address 10.200.20.12/24, gateway 10.200.20.1 acquired from 168.63.129.16 Apr 30 00:36:19.022508 ignition[788]: Ignition 2.19.0 Apr 30 00:36:19.022519 ignition[788]: Stage: fetch-offline Apr 30 00:36:19.027382 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 30 00:36:19.022560 ignition[788]: no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:19.022568 ignition[788]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:19.022665 ignition[788]: parsed url from cmdline: "" Apr 30 00:36:19.022668 ignition[788]: no config URL provided Apr 30 00:36:19.022672 ignition[788]: reading system config file "/usr/lib/ignition/user.ign" Apr 30 00:36:19.053375 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Apr 30 00:36:19.022678 ignition[788]: no config at "/usr/lib/ignition/user.ign" Apr 30 00:36:19.022683 ignition[788]: failed to fetch config: resource requires networking Apr 30 00:36:19.022869 ignition[788]: Ignition finished successfully Apr 30 00:36:19.073333 ignition[882]: Ignition 2.19.0 Apr 30 00:36:19.073340 ignition[882]: Stage: fetch Apr 30 00:36:19.073543 ignition[882]: no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:19.073554 ignition[882]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:19.073661 ignition[882]: parsed url from cmdline: "" Apr 30 00:36:19.073664 ignition[882]: no config URL provided Apr 30 00:36:19.073668 ignition[882]: reading system config file "/usr/lib/ignition/user.ign" Apr 30 00:36:19.073676 ignition[882]: no config at "/usr/lib/ignition/user.ign" Apr 30 00:36:19.073707 ignition[882]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Apr 30 00:36:19.185421 ignition[882]: GET result: OK Apr 30 00:36:19.185541 ignition[882]: config has been read from IMDS userdata Apr 30 00:36:19.185581 ignition[882]: parsing config with SHA512: c74a6cc6e4182038b30a5bb84c7ceca9605595d73dfdb51b912378251a996c0afd24e625706b4d4b5a3447b7f04c1949bc5afeef0132bd898fd6fb0ae5d80175 Apr 30 00:36:19.189590 unknown[882]: fetched base config from "system" Apr 30 00:36:19.189986 ignition[882]: fetch: fetch complete Apr 30 00:36:19.189597 unknown[882]: fetched base config from "system" Apr 30 00:36:19.189991 ignition[882]: fetch: fetch passed Apr 30 00:36:19.189602 unknown[882]: fetched user config from "azure" Apr 30 00:36:19.190031 ignition[882]: Ignition finished successfully Apr 30 00:36:19.192202 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Apr 30 00:36:19.215373 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 30 00:36:19.242955 ignition[890]: Ignition 2.19.0 Apr 30 00:36:19.242970 ignition[890]: Stage: kargs Apr 30 00:36:19.243185 ignition[890]: no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:19.250324 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 30 00:36:19.243198 ignition[890]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:19.244161 ignition[890]: kargs: kargs passed Apr 30 00:36:19.244210 ignition[890]: Ignition finished successfully Apr 30 00:36:19.277364 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 30 00:36:19.295718 ignition[896]: Ignition 2.19.0 Apr 30 00:36:19.295730 ignition[896]: Stage: disks Apr 30 00:36:19.301127 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 30 00:36:19.295901 ignition[896]: no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:19.308754 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 30 00:36:19.295910 ignition[896]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:19.319219 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 30 00:36:19.296845 ignition[896]: disks: disks passed Apr 30 00:36:19.330355 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 30 00:36:19.296895 ignition[896]: Ignition finished successfully Apr 30 00:36:19.341783 systemd[1]: Reached target sysinit.target - System Initialization. Apr 30 00:36:19.354928 systemd[1]: Reached target basic.target - Basic System. Apr 30 00:36:19.367029 systemd-networkd[872]: eth0: Gained IPv6LL Apr 30 00:36:19.367391 systemd-networkd[872]: enP31290s1: Gained IPv6LL Apr 30 00:36:19.384354 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 30 00:36:19.477835 systemd-fsck[904]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Apr 30 00:36:19.487211 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 30 00:36:19.504354 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 30 00:36:19.567091 kernel: EXT4-fs (sda9): mounted filesystem c13301f3-70ec-4948-963a-f1db0e953273 r/w with ordered data mode. Quota mode: none. Apr 30 00:36:19.567476 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 30 00:36:19.572740 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 30 00:36:19.625188 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 30 00:36:19.632238 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 30 00:36:19.643250 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Apr 30 00:36:19.657563 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 30 00:36:19.657599 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 30 00:36:19.665570 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 30 00:36:19.707262 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (915) Apr 30 00:36:19.707322 kernel: BTRFS info (device sda6): first mount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:19.707846 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 30 00:36:19.729716 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 30 00:36:19.729741 kernel: BTRFS info (device sda6): using free space tree Apr 30 00:36:19.736142 kernel: BTRFS info (device sda6): auto enabling async discard Apr 30 00:36:19.736981 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 30 00:36:20.212952 coreos-metadata[917]: Apr 30 00:36:20.212 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Apr 30 00:36:20.222900 coreos-metadata[917]: Apr 30 00:36:20.222 INFO Fetch successful Apr 30 00:36:20.222900 coreos-metadata[917]: Apr 30 00:36:20.222 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Apr 30 00:36:20.239426 coreos-metadata[917]: Apr 30 00:36:20.234 INFO Fetch successful Apr 30 00:36:20.239426 coreos-metadata[917]: Apr 30 00:36:20.235 INFO wrote hostname ci-4081.3.3-a-c388cdac8a to /sysroot/etc/hostname Apr 30 00:36:20.239809 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 30 00:36:20.547677 initrd-setup-root[945]: cut: /sysroot/etc/passwd: No such file or directory Apr 30 00:36:20.625038 initrd-setup-root[952]: cut: /sysroot/etc/group: No such file or directory Apr 30 00:36:20.647548 initrd-setup-root[959]: cut: /sysroot/etc/shadow: No such file or directory Apr 30 00:36:20.657606 initrd-setup-root[966]: cut: /sysroot/etc/gshadow: No such file or directory Apr 30 00:36:21.426933 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 30 00:36:21.440570 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 30 00:36:21.450328 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 30 00:36:21.472188 kernel: BTRFS info (device sda6): last unmount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:21.466805 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 30 00:36:21.496708 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 30 00:36:21.511096 ignition[1035]: INFO : Ignition 2.19.0 Apr 30 00:36:21.511096 ignition[1035]: INFO : Stage: mount Apr 30 00:36:21.525274 ignition[1035]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:21.525274 ignition[1035]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:21.525274 ignition[1035]: INFO : mount: mount passed Apr 30 00:36:21.525274 ignition[1035]: INFO : Ignition finished successfully Apr 30 00:36:21.516269 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 30 00:36:21.542345 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 30 00:36:21.561285 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 30 00:36:21.601774 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (1045) Apr 30 00:36:21.601847 kernel: BTRFS info (device sda6): first mount of filesystem ece78588-c2c6-41f3-bdc2-614da63113c1 Apr 30 00:36:21.608567 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 30 00:36:21.613350 kernel: BTRFS info (device sda6): using free space tree Apr 30 00:36:21.620096 kernel: BTRFS info (device sda6): auto enabling async discard Apr 30 00:36:21.621224 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 30 00:36:21.651254 ignition[1062]: INFO : Ignition 2.19.0 Apr 30 00:36:21.651254 ignition[1062]: INFO : Stage: files Apr 30 00:36:21.659467 ignition[1062]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:21.659467 ignition[1062]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:21.659467 ignition[1062]: DEBUG : files: compiled without relabeling support, skipping Apr 30 00:36:21.682558 ignition[1062]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 30 00:36:21.682558 ignition[1062]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 30 00:36:21.771328 ignition[1062]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 30 00:36:21.779875 ignition[1062]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 30 00:36:21.779875 ignition[1062]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 30 00:36:21.772467 unknown[1062]: wrote ssh authorized keys file for user: core Apr 30 00:36:21.808837 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 30 00:36:21.819274 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Apr 30 00:36:21.899008 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 30 00:36:22.002707 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 30 00:36:22.002707 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 00:36:22.022078 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-arm64.raw: attempt #1 Apr 30 00:36:22.429357 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Apr 30 00:36:22.689965 ignition[1062]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 00:36:22.689965 ignition[1062]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Apr 30 00:36:22.736102 ignition[1062]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 30 00:36:22.748075 ignition[1062]: INFO : files: files passed Apr 30 00:36:22.748075 ignition[1062]: INFO : Ignition finished successfully Apr 30 00:36:22.760109 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 30 00:36:22.792392 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 30 00:36:22.811289 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 30 00:36:22.835869 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 30 00:36:22.835976 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 30 00:36:22.885589 initrd-setup-root-after-ignition[1095]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 30 00:36:22.893700 initrd-setup-root-after-ignition[1091]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 30 00:36:22.893700 initrd-setup-root-after-ignition[1091]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 30 00:36:22.885883 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 30 00:36:22.900963 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 30 00:36:22.941424 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 30 00:36:22.976311 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 30 00:36:22.976432 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 30 00:36:22.989183 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 30 00:36:23.000858 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 30 00:36:23.011516 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 30 00:36:23.029362 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 30 00:36:23.051582 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 30 00:36:23.068312 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 30 00:36:23.086593 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 30 00:36:23.093686 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 00:36:23.107426 systemd[1]: Stopped target timers.target - Timer Units. Apr 30 00:36:23.118935 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 30 00:36:23.119140 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 30 00:36:23.135612 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 30 00:36:23.147465 systemd[1]: Stopped target basic.target - Basic System. Apr 30 00:36:23.160313 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 30 00:36:23.171642 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 30 00:36:23.188251 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 30 00:36:23.200533 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 30 00:36:23.211549 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 30 00:36:23.223492 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 30 00:36:23.235683 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 30 00:36:23.246442 systemd[1]: Stopped target swap.target - Swaps. Apr 30 00:36:23.256283 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 30 00:36:23.256466 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 30 00:36:23.274955 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 30 00:36:23.286198 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 00:36:23.298150 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 30 00:36:23.304023 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 00:36:23.310802 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 30 00:36:23.310998 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 30 00:36:23.328635 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 30 00:36:23.328826 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 30 00:36:23.340119 systemd[1]: ignition-files.service: Deactivated successfully. Apr 30 00:36:23.340276 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 30 00:36:23.350327 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Apr 30 00:36:23.350487 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 30 00:36:23.381238 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 30 00:36:23.409184 ignition[1115]: INFO : Ignition 2.19.0 Apr 30 00:36:23.409184 ignition[1115]: INFO : Stage: umount Apr 30 00:36:23.409184 ignition[1115]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 00:36:23.409184 ignition[1115]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 30 00:36:23.409184 ignition[1115]: INFO : umount: umount passed Apr 30 00:36:23.409184 ignition[1115]: INFO : Ignition finished successfully Apr 30 00:36:23.390010 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 30 00:36:23.390320 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 00:36:23.429236 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 30 00:36:23.438884 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 30 00:36:23.439461 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 00:36:23.454305 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 30 00:36:23.454483 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 30 00:36:23.473867 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 30 00:36:23.473963 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 30 00:36:23.486327 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 30 00:36:23.486430 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 30 00:36:23.495305 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 30 00:36:23.495362 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 30 00:36:23.513267 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 30 00:36:23.513340 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 30 00:36:23.524492 systemd[1]: ignition-fetch.service: Deactivated successfully. Apr 30 00:36:23.524560 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Apr 30 00:36:23.530887 systemd[1]: Stopped target network.target - Network. Apr 30 00:36:23.540840 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 30 00:36:23.540919 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 30 00:36:23.551778 systemd[1]: Stopped target paths.target - Path Units. Apr 30 00:36:23.562107 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 30 00:36:23.566949 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 00:36:23.574113 systemd[1]: Stopped target slices.target - Slice Units. Apr 30 00:36:23.585641 systemd[1]: Stopped target sockets.target - Socket Units. Apr 30 00:36:23.595632 systemd[1]: iscsid.socket: Deactivated successfully. Apr 30 00:36:23.595696 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 30 00:36:23.605309 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 30 00:36:23.605354 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 30 00:36:23.617301 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 30 00:36:23.617370 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 30 00:36:23.627836 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 30 00:36:23.627899 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 30 00:36:23.638157 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 30 00:36:23.851159 kernel: hv_netvsc 002248b7-44a6-0022-48b7-44a6002248b7 eth0: Data path switched from VF: enP31290s1 Apr 30 00:36:23.648785 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 30 00:36:23.660255 systemd-networkd[872]: eth0: DHCPv6 lease lost Apr 30 00:36:23.662486 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 30 00:36:23.663844 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 30 00:36:23.664269 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 30 00:36:23.674997 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 30 00:36:23.675037 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 30 00:36:23.699205 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 30 00:36:23.710839 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 30 00:36:23.710923 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 30 00:36:23.723583 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 00:36:23.739859 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 30 00:36:23.739986 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 30 00:36:23.757229 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 30 00:36:23.757377 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 30 00:36:23.768427 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 30 00:36:23.768497 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 30 00:36:23.774719 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 30 00:36:23.774773 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 00:36:23.788529 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 30 00:36:23.790115 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 00:36:23.801557 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 30 00:36:23.801640 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 30 00:36:23.812876 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 30 00:36:23.812925 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 00:36:23.824459 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 30 00:36:23.824523 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 30 00:36:23.846248 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 30 00:36:23.846363 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 30 00:36:23.864823 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 30 00:36:23.864892 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 00:36:23.890377 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 30 00:36:23.904256 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 30 00:36:23.904349 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 00:36:23.928892 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 00:36:23.928962 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:23.942128 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 30 00:36:23.942248 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 30 00:36:24.002227 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 30 00:36:24.002355 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 30 00:36:24.119914 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 30 00:36:24.120116 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 30 00:36:24.126528 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 30 00:36:24.138040 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 30 00:36:24.138125 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 30 00:36:24.166397 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 30 00:36:24.185638 systemd[1]: Switching root. Apr 30 00:36:24.290659 systemd-journald[217]: Journal stopped Apr 30 00:36:28.721566 systemd-journald[217]: Received SIGTERM from PID 1 (systemd). Apr 30 00:36:28.721590 kernel: SELinux: policy capability network_peer_controls=1 Apr 30 00:36:28.721600 kernel: SELinux: policy capability open_perms=1 Apr 30 00:36:28.721610 kernel: SELinux: policy capability extended_socket_class=1 Apr 30 00:36:28.721618 kernel: SELinux: policy capability always_check_network=0 Apr 30 00:36:28.721625 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 30 00:36:28.721636 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 30 00:36:28.721644 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 30 00:36:28.721652 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 30 00:36:28.721660 kernel: audit: type=1403 audit(1745973385.353:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 30 00:36:28.721670 systemd[1]: Successfully loaded SELinux policy in 232.111ms. Apr 30 00:36:28.721680 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 10.870ms. Apr 30 00:36:28.721689 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 30 00:36:28.721698 systemd[1]: Detected virtualization microsoft. Apr 30 00:36:28.721707 systemd[1]: Detected architecture arm64. Apr 30 00:36:28.721717 systemd[1]: Detected first boot. Apr 30 00:36:28.721726 systemd[1]: Hostname set to . Apr 30 00:36:28.721735 systemd[1]: Initializing machine ID from random generator. Apr 30 00:36:28.721744 zram_generator::config[1157]: No configuration found. Apr 30 00:36:28.721753 systemd[1]: Populated /etc with preset unit settings. Apr 30 00:36:28.721763 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 30 00:36:28.721773 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Apr 30 00:36:28.721782 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 30 00:36:28.721792 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 30 00:36:28.721801 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 30 00:36:28.721810 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 30 00:36:28.721819 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 30 00:36:28.721829 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 30 00:36:28.721840 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 30 00:36:28.721850 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 30 00:36:28.721859 systemd[1]: Created slice user.slice - User and Session Slice. Apr 30 00:36:28.721868 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 00:36:28.721877 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 00:36:28.721886 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 30 00:36:28.721896 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 30 00:36:28.721905 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 30 00:36:28.721914 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 30 00:36:28.721925 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Apr 30 00:36:28.721934 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 00:36:28.721943 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Apr 30 00:36:28.721954 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Apr 30 00:36:28.721964 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Apr 30 00:36:28.721974 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 30 00:36:28.721983 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 00:36:28.721994 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 30 00:36:28.722003 systemd[1]: Reached target slices.target - Slice Units. Apr 30 00:36:28.722013 systemd[1]: Reached target swap.target - Swaps. Apr 30 00:36:28.722022 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 30 00:36:28.722032 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 30 00:36:28.722042 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 30 00:36:28.722051 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 30 00:36:28.722063 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 00:36:28.722088 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 30 00:36:28.722098 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 30 00:36:28.722107 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 30 00:36:28.722117 systemd[1]: Mounting media.mount - External Media Directory... Apr 30 00:36:28.722126 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 30 00:36:28.722139 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 30 00:36:28.722149 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 30 00:36:28.722159 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 30 00:36:28.722168 systemd[1]: Reached target machines.target - Containers. Apr 30 00:36:28.722178 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 30 00:36:28.722187 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 00:36:28.722197 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 30 00:36:28.722206 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 30 00:36:28.722217 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 30 00:36:28.722227 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 30 00:36:28.722236 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 30 00:36:28.722246 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 30 00:36:28.722256 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 30 00:36:28.722266 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 30 00:36:28.722276 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 30 00:36:28.722286 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Apr 30 00:36:28.722295 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 30 00:36:28.722306 systemd[1]: Stopped systemd-fsck-usr.service. Apr 30 00:36:28.722315 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 30 00:36:28.722324 kernel: loop: module loaded Apr 30 00:36:28.722333 kernel: fuse: init (API version 7.39) Apr 30 00:36:28.722341 kernel: ACPI: bus type drm_connector registered Apr 30 00:36:28.722350 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 30 00:36:28.722360 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 30 00:36:28.722369 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 30 00:36:28.722397 systemd-journald[1260]: Collecting audit messages is disabled. Apr 30 00:36:28.722419 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 30 00:36:28.722430 systemd-journald[1260]: Journal started Apr 30 00:36:28.722453 systemd-journald[1260]: Runtime Journal (/run/log/journal/747b625e75d54be9a2d466fa73813828) is 8.0M, max 78.5M, 70.5M free. Apr 30 00:36:27.586391 systemd[1]: Queued start job for default target multi-user.target. Apr 30 00:36:27.732926 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Apr 30 00:36:27.733342 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 30 00:36:27.733656 systemd[1]: systemd-journald.service: Consumed 3.214s CPU time. Apr 30 00:36:28.739092 systemd[1]: verity-setup.service: Deactivated successfully. Apr 30 00:36:28.739167 systemd[1]: Stopped verity-setup.service. Apr 30 00:36:28.758895 systemd[1]: Started systemd-journald.service - Journal Service. Apr 30 00:36:28.759771 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 30 00:36:28.765821 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 30 00:36:28.771833 systemd[1]: Mounted media.mount - External Media Directory. Apr 30 00:36:28.777188 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 30 00:36:28.783550 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 30 00:36:28.789376 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 30 00:36:28.794617 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 30 00:36:28.802658 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 00:36:28.809688 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 30 00:36:28.809848 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 30 00:36:28.816623 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 30 00:36:28.816760 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 30 00:36:28.823044 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 30 00:36:28.823196 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 30 00:36:28.828970 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 30 00:36:28.829116 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 30 00:36:28.836020 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 30 00:36:28.836178 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 30 00:36:28.842339 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 30 00:36:28.842476 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 30 00:36:28.848907 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 30 00:36:28.854982 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 30 00:36:28.861971 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 30 00:36:28.868868 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 00:36:28.884510 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 30 00:36:28.897169 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 30 00:36:28.904175 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 30 00:36:28.910562 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 30 00:36:28.910600 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 30 00:36:28.917334 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Apr 30 00:36:28.925435 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 30 00:36:28.932512 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 30 00:36:28.937777 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 00:36:29.041287 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 30 00:36:29.048744 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 30 00:36:29.055274 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 30 00:36:29.058305 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 30 00:36:29.068439 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 30 00:36:29.069580 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 30 00:36:29.079326 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 30 00:36:29.095299 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 30 00:36:29.104347 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 30 00:36:29.114235 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 30 00:36:29.121394 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 30 00:36:29.131796 systemd-journald[1260]: Time spent on flushing to /var/log/journal/747b625e75d54be9a2d466fa73813828 is 91.276ms for 899 entries. Apr 30 00:36:29.131796 systemd-journald[1260]: System Journal (/var/log/journal/747b625e75d54be9a2d466fa73813828) is 11.8M, max 2.6G, 2.6G free. Apr 30 00:36:29.345874 systemd-journald[1260]: Received client request to flush runtime journal. Apr 30 00:36:29.346223 systemd-journald[1260]: /var/log/journal/747b625e75d54be9a2d466fa73813828/system.journal: Realtime clock jumped backwards relative to last journal entry, rotating. Apr 30 00:36:29.346317 kernel: loop0: detected capacity change from 0 to 31320 Apr 30 00:36:29.346342 systemd-journald[1260]: Rotating system journal. Apr 30 00:36:29.139664 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 30 00:36:29.147954 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 30 00:36:29.157973 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 30 00:36:29.190503 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Apr 30 00:36:29.198426 udevadm[1294]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Apr 30 00:36:29.201504 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 30 00:36:29.321488 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 30 00:36:29.335353 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 30 00:36:29.348582 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 30 00:36:29.386824 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 30 00:36:29.388126 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Apr 30 00:36:29.501965 systemd-tmpfiles[1307]: ACLs are not supported, ignoring. Apr 30 00:36:29.501980 systemd-tmpfiles[1307]: ACLs are not supported, ignoring. Apr 30 00:36:29.506540 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 00:36:29.657096 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 30 00:36:29.710099 kernel: loop1: detected capacity change from 0 to 189592 Apr 30 00:36:29.792269 kernel: loop2: detected capacity change from 0 to 114432 Apr 30 00:36:30.176101 kernel: loop3: detected capacity change from 0 to 114328 Apr 30 00:36:30.530108 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 30 00:36:30.541270 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 00:36:30.569083 systemd-udevd[1317]: Using default interface naming scheme 'v255'. Apr 30 00:36:30.595856 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 00:36:30.613270 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 30 00:36:30.635123 kernel: loop4: detected capacity change from 0 to 31320 Apr 30 00:36:30.653098 kernel: loop5: detected capacity change from 0 to 189592 Apr 30 00:36:30.656307 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Apr 30 00:36:30.679098 kernel: loop6: detected capacity change from 0 to 114432 Apr 30 00:36:30.689480 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 30 00:36:30.692093 kernel: loop7: detected capacity change from 0 to 114328 Apr 30 00:36:30.701646 (sd-merge)[1333]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Apr 30 00:36:30.702457 (sd-merge)[1333]: Merged extensions into '/usr'. Apr 30 00:36:30.718538 systemd[1]: Reloading requested from client PID 1291 ('systemd-sysext') (unit systemd-sysext.service)... Apr 30 00:36:30.718561 systemd[1]: Reloading... Apr 30 00:36:30.830837 zram_generator::config[1392]: No configuration found. Apr 30 00:36:30.854188 kernel: mousedev: PS/2 mouse device common for all mice Apr 30 00:36:30.939130 kernel: hv_vmbus: registering driver hv_balloon Apr 30 00:36:30.939222 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Apr 30 00:36:30.943779 kernel: hv_balloon: Memory hot add disabled on ARM64 Apr 30 00:36:30.948424 kernel: hv_vmbus: registering driver hyperv_fb Apr 30 00:36:30.953370 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Apr 30 00:36:30.962127 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Apr 30 00:36:30.967084 kernel: Console: switching to colour dummy device 80x25 Apr 30 00:36:30.975146 kernel: Console: switching to colour frame buffer device 128x48 Apr 30 00:36:30.991807 systemd-networkd[1329]: lo: Link UP Apr 30 00:36:30.991819 systemd-networkd[1329]: lo: Gained carrier Apr 30 00:36:30.995224 systemd-networkd[1329]: Enumeration completed Apr 30 00:36:30.996262 systemd-networkd[1329]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 00:36:30.996329 systemd-networkd[1329]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 00:36:31.012108 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (1332) Apr 30 00:36:31.062104 kernel: mlx5_core 7a3a:00:02.0 enP31290s1: Link up Apr 30 00:36:31.093391 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 00:36:31.103314 kernel: hv_netvsc 002248b7-44a6-0022-48b7-44a6002248b7 eth0: Data path switched to VF: enP31290s1 Apr 30 00:36:31.105386 systemd-networkd[1329]: enP31290s1: Link UP Apr 30 00:36:31.105476 systemd-networkd[1329]: eth0: Link UP Apr 30 00:36:31.105480 systemd-networkd[1329]: eth0: Gained carrier Apr 30 00:36:31.105495 systemd-networkd[1329]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 00:36:31.109398 systemd-networkd[1329]: enP31290s1: Gained carrier Apr 30 00:36:31.115126 systemd-networkd[1329]: eth0: DHCPv4 address 10.200.20.12/24, gateway 10.200.20.1 acquired from 168.63.129.16 Apr 30 00:36:31.170010 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Apr 30 00:36:31.176273 systemd[1]: Reloading finished in 457 ms. Apr 30 00:36:31.204650 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 30 00:36:31.210779 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 30 00:36:31.218722 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 30 00:36:31.257369 systemd[1]: Starting ensure-sysext.service... Apr 30 00:36:31.262365 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 30 00:36:31.270509 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 30 00:36:31.286222 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 30 00:36:31.295309 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 00:36:31.306444 systemd-tmpfiles[1485]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 30 00:36:31.308584 systemd[1]: Reloading requested from client PID 1482 ('systemctl') (unit ensure-sysext.service)... Apr 30 00:36:31.308602 systemd[1]: Reloading... Apr 30 00:36:31.309467 systemd-tmpfiles[1485]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 30 00:36:31.310173 systemd-tmpfiles[1485]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 30 00:36:31.310400 systemd-tmpfiles[1485]: ACLs are not supported, ignoring. Apr 30 00:36:31.310444 systemd-tmpfiles[1485]: ACLs are not supported, ignoring. Apr 30 00:36:31.333894 systemd-tmpfiles[1485]: Detected autofs mount point /boot during canonicalization of boot. Apr 30 00:36:31.333908 systemd-tmpfiles[1485]: Skipping /boot Apr 30 00:36:31.350522 systemd-tmpfiles[1485]: Detected autofs mount point /boot during canonicalization of boot. Apr 30 00:36:31.350538 systemd-tmpfiles[1485]: Skipping /boot Apr 30 00:36:31.402099 zram_generator::config[1521]: No configuration found. Apr 30 00:36:31.519541 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 00:36:31.598160 systemd[1]: Reloading finished in 289 ms. Apr 30 00:36:31.617922 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Apr 30 00:36:31.630676 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 30 00:36:31.637920 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 00:36:31.660026 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 30 00:36:31.670220 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 30 00:36:31.679471 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Apr 30 00:36:31.693697 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 30 00:36:31.703230 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 30 00:36:31.712224 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 30 00:36:31.723362 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 00:36:31.728194 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 30 00:36:31.744577 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 30 00:36:31.754333 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 30 00:36:31.761790 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 00:36:31.762719 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 30 00:36:31.764106 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 30 00:36:31.785195 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 00:36:31.789444 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 30 00:36:31.800823 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 00:36:31.802343 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 30 00:36:31.812619 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 30 00:36:31.812790 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 30 00:36:31.817088 lvm[1589]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 30 00:36:31.822031 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 30 00:36:31.822471 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 30 00:36:31.828967 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 30 00:36:31.829191 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 30 00:36:31.845928 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Apr 30 00:36:31.853362 systemd-resolved[1591]: Positive Trust Anchors: Apr 30 00:36:31.853377 systemd-resolved[1591]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 30 00:36:31.853409 systemd-resolved[1591]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 30 00:36:31.856227 augenrules[1611]: No rules Apr 30 00:36:31.857433 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 30 00:36:31.863690 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 00:36:31.869446 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Apr 30 00:36:31.878907 systemd-resolved[1591]: Using system hostname 'ci-4081.3.3-a-c388cdac8a'. Apr 30 00:36:31.885447 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 30 00:36:31.892723 lvm[1618]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 30 00:36:31.903434 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 30 00:36:31.910796 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 30 00:36:31.920312 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 30 00:36:31.927643 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 00:36:31.928268 systemd[1]: Reached target time-set.target - System Time Set. Apr 30 00:36:31.937174 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 30 00:36:31.944436 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 00:36:31.953916 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 30 00:36:31.961151 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Apr 30 00:36:31.969677 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 30 00:36:31.969981 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 30 00:36:31.978720 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 30 00:36:31.979033 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 30 00:36:31.986283 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 30 00:36:31.988112 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 30 00:36:31.996851 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 30 00:36:31.997000 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 30 00:36:32.005539 systemd[1]: Finished ensure-sysext.service. Apr 30 00:36:32.015389 systemd[1]: Reached target network.target - Network. Apr 30 00:36:32.020603 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 30 00:36:32.026787 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 30 00:36:32.026864 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 30 00:36:32.036681 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 30 00:36:32.271489 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 30 00:36:32.278979 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 30 00:36:32.418227 systemd-networkd[1329]: eth0: Gained IPv6LL Apr 30 00:36:32.421023 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 30 00:36:32.429204 systemd[1]: Reached target network-online.target - Network is Online. Apr 30 00:36:32.482220 systemd-networkd[1329]: enP31290s1: Gained IPv6LL Apr 30 00:36:35.286398 ldconfig[1286]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 30 00:36:35.301719 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 30 00:36:35.312247 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 30 00:36:35.332317 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 30 00:36:35.338654 systemd[1]: Reached target sysinit.target - System Initialization. Apr 30 00:36:35.344520 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 30 00:36:35.351382 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 30 00:36:35.358053 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 30 00:36:35.363827 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 30 00:36:35.370343 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 30 00:36:35.376780 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 30 00:36:35.376822 systemd[1]: Reached target paths.target - Path Units. Apr 30 00:36:35.381999 systemd[1]: Reached target timers.target - Timer Units. Apr 30 00:36:35.419673 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 30 00:36:35.427046 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 30 00:36:35.437288 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 30 00:36:35.443410 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 30 00:36:35.449233 systemd[1]: Reached target sockets.target - Socket Units. Apr 30 00:36:35.454103 systemd[1]: Reached target basic.target - Basic System. Apr 30 00:36:35.459609 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 30 00:36:35.459644 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 30 00:36:35.467191 systemd[1]: Starting chronyd.service - NTP client/server... Apr 30 00:36:35.475255 systemd[1]: Starting containerd.service - containerd container runtime... Apr 30 00:36:35.486325 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Apr 30 00:36:35.492941 (chronyd)[1640]: chronyd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Apr 30 00:36:35.496701 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 30 00:36:35.506279 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 30 00:36:35.525813 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 30 00:36:35.532605 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 30 00:36:35.532961 systemd[1]: hv_fcopy_daemon.service - Hyper-V FCOPY daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_fcopy). Apr 30 00:36:35.536334 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. Apr 30 00:36:35.542800 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). Apr 30 00:36:35.545103 KVP[1648]: KVP starting; pid is:1648 Apr 30 00:36:35.545261 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:36:35.553732 chronyd[1652]: chronyd version 4.5 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +ASYNCDNS +NTS +SECHASH +IPV6 -DEBUG) Apr 30 00:36:35.557117 jq[1644]: false Apr 30 00:36:35.566291 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 30 00:36:35.573409 chronyd[1652]: Timezone right/UTC failed leap second check, ignoring Apr 30 00:36:35.573629 chronyd[1652]: Loaded seccomp filter (level 2) Apr 30 00:36:35.577826 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 30 00:36:35.585268 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 30 00:36:35.593305 extend-filesystems[1647]: Found loop4 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found loop5 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found loop6 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found loop7 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found sda Apr 30 00:36:35.593305 extend-filesystems[1647]: Found sda1 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found sda2 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found sda3 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found usr Apr 30 00:36:35.593305 extend-filesystems[1647]: Found sda4 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found sda6 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found sda7 Apr 30 00:36:35.593305 extend-filesystems[1647]: Found sda9 Apr 30 00:36:35.593305 extend-filesystems[1647]: Checking size of /dev/sda9 Apr 30 00:36:35.732111 kernel: hv_utils: KVP IC version 4.0 Apr 30 00:36:35.655096 KVP[1648]: KVP LIC Version: 3.1 Apr 30 00:36:35.596323 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 30 00:36:35.732359 extend-filesystems[1647]: Old size kept for /dev/sda9 Apr 30 00:36:35.732359 extend-filesystems[1647]: Found sr0 Apr 30 00:36:35.717595 dbus-daemon[1643]: [system] SELinux support is enabled Apr 30 00:36:35.621311 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 30 00:36:35.650249 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 30 00:36:35.665680 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 30 00:36:35.666225 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 30 00:36:35.675373 systemd[1]: Starting update-engine.service - Update Engine... Apr 30 00:36:35.699231 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 30 00:36:35.747603 jq[1676]: true Apr 30 00:36:35.716261 systemd[1]: Started chronyd.service - NTP client/server. Apr 30 00:36:35.727828 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 30 00:36:35.743611 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 30 00:36:35.743792 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 30 00:36:35.744088 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 30 00:36:35.745136 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 30 00:36:35.767535 systemd[1]: motdgen.service: Deactivated successfully. Apr 30 00:36:35.768155 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 30 00:36:35.788952 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 30 00:36:35.797187 systemd-logind[1668]: Watching system buttons on /dev/input/event1 (AT Translated Set 2 keyboard) Apr 30 00:36:35.800386 systemd-logind[1668]: New seat seat0. Apr 30 00:36:35.800610 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 30 00:36:35.800819 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 30 00:36:35.829181 update_engine[1671]: I20250430 00:36:35.829047 1671 main.cc:92] Flatcar Update Engine starting Apr 30 00:36:35.841104 update_engine[1671]: I20250430 00:36:35.833174 1671 update_check_scheduler.cc:74] Next update check in 7m49s Apr 30 00:36:35.834428 systemd[1]: Started systemd-logind.service - User Login Management. Apr 30 00:36:35.850603 jq[1697]: true Apr 30 00:36:35.865333 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 30 00:36:35.865786 dbus-daemon[1643]: [system] Successfully activated service 'org.freedesktop.systemd1' Apr 30 00:36:35.865401 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 30 00:36:35.875594 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 30 00:36:35.875639 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 30 00:36:35.876149 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (1690) Apr 30 00:36:35.884376 (ntainerd)[1704]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 30 00:36:35.892528 systemd[1]: Started update-engine.service - Update Engine. Apr 30 00:36:35.901815 coreos-metadata[1642]: Apr 30 00:36:35.900 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Apr 30 00:36:35.908607 coreos-metadata[1642]: Apr 30 00:36:35.907 INFO Fetch successful Apr 30 00:36:35.908607 coreos-metadata[1642]: Apr 30 00:36:35.907 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Apr 30 00:36:35.910949 coreos-metadata[1642]: Apr 30 00:36:35.910 INFO Fetch successful Apr 30 00:36:35.910949 coreos-metadata[1642]: Apr 30 00:36:35.910 INFO Fetching http://168.63.129.16/machine/1178be7f-7415-473e-a9d4-6667daf3eff6/87aee4c4%2D58a6%2D45e3%2D953e%2Defc21f8b97c9.%5Fci%2D4081.3.3%2Da%2Dc388cdac8a?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Apr 30 00:36:35.915800 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 30 00:36:35.924656 coreos-metadata[1642]: Apr 30 00:36:35.924 INFO Fetch successful Apr 30 00:36:35.924656 coreos-metadata[1642]: Apr 30 00:36:35.924 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Apr 30 00:36:35.925227 tar[1695]: linux-arm64/helm Apr 30 00:36:35.935476 coreos-metadata[1642]: Apr 30 00:36:35.935 INFO Fetch successful Apr 30 00:36:36.009129 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Apr 30 00:36:36.016842 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 30 00:36:36.128983 locksmithd[1722]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 30 00:36:36.147741 bash[1749]: Updated "/home/core/.ssh/authorized_keys" Apr 30 00:36:36.149506 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 30 00:36:36.160336 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Apr 30 00:36:36.450184 tar[1695]: linux-arm64/LICENSE Apr 30 00:36:36.450184 tar[1695]: linux-arm64/README.md Apr 30 00:36:36.471274 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 30 00:36:36.749623 containerd[1704]: time="2025-04-30T00:36:36.749445520Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Apr 30 00:36:36.779375 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:36:36.787755 (kubelet)[1780]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 00:36:36.819510 containerd[1704]: time="2025-04-30T00:36:36.819455120Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 30 00:36:36.823421 containerd[1704]: time="2025-04-30T00:36:36.823368000Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.88-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 30 00:36:36.823870 containerd[1704]: time="2025-04-30T00:36:36.823850640Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 30 00:36:36.823942 containerd[1704]: time="2025-04-30T00:36:36.823928920Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 30 00:36:36.824201 containerd[1704]: time="2025-04-30T00:36:36.824181600Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Apr 30 00:36:36.825089 containerd[1704]: time="2025-04-30T00:36:36.824271920Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Apr 30 00:36:36.825152 containerd[1704]: time="2025-04-30T00:36:36.825052000Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 00:36:36.825218 containerd[1704]: time="2025-04-30T00:36:36.825204640Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 30 00:36:36.825498 containerd[1704]: time="2025-04-30T00:36:36.825477000Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 00:36:36.825744 containerd[1704]: time="2025-04-30T00:36:36.825728160Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 30 00:36:36.825807 containerd[1704]: time="2025-04-30T00:36:36.825793600Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 00:36:36.825853 containerd[1704]: time="2025-04-30T00:36:36.825841560Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 30 00:36:36.826822 containerd[1704]: time="2025-04-30T00:36:36.826541000Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 30 00:36:36.826822 containerd[1704]: time="2025-04-30T00:36:36.826787080Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 30 00:36:36.827061 containerd[1704]: time="2025-04-30T00:36:36.827041560Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 00:36:36.827572 containerd[1704]: time="2025-04-30T00:36:36.827554440Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 30 00:36:36.827734 containerd[1704]: time="2025-04-30T00:36:36.827717160Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 30 00:36:36.828383 containerd[1704]: time="2025-04-30T00:36:36.828250880Z" level=info msg="metadata content store policy set" policy=shared Apr 30 00:36:36.843990 containerd[1704]: time="2025-04-30T00:36:36.843944640Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 30 00:36:36.844197 containerd[1704]: time="2025-04-30T00:36:36.844174680Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 30 00:36:36.844318 containerd[1704]: time="2025-04-30T00:36:36.844301600Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Apr 30 00:36:36.845317 containerd[1704]: time="2025-04-30T00:36:36.844989200Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Apr 30 00:36:36.845317 containerd[1704]: time="2025-04-30T00:36:36.845016240Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 30 00:36:36.845317 containerd[1704]: time="2025-04-30T00:36:36.845265560Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 30 00:36:36.846477 containerd[1704]: time="2025-04-30T00:36:36.846230280Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 30 00:36:36.846477 containerd[1704]: time="2025-04-30T00:36:36.846412560Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Apr 30 00:36:36.846477 containerd[1704]: time="2025-04-30T00:36:36.846432960Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Apr 30 00:36:36.846477 containerd[1704]: time="2025-04-30T00:36:36.846447840Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Apr 30 00:36:36.846611 containerd[1704]: time="2025-04-30T00:36:36.846461680Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 30 00:36:36.846678 containerd[1704]: time="2025-04-30T00:36:36.846655480Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.846786600Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.846809120Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.847336480Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.847363040Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.847377040Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.847389800Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.847423920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.847440560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.847453680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847491 containerd[1704]: time="2025-04-30T00:36:36.847472800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847803 containerd[1704]: time="2025-04-30T00:36:36.847735120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847803 containerd[1704]: time="2025-04-30T00:36:36.847758680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847803 containerd[1704]: time="2025-04-30T00:36:36.847771120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847961 containerd[1704]: time="2025-04-30T00:36:36.847788760Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847961 containerd[1704]: time="2025-04-30T00:36:36.847899440Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847961 containerd[1704]: time="2025-04-30T00:36:36.847917520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847961 containerd[1704]: time="2025-04-30T00:36:36.847929000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.847961 containerd[1704]: time="2025-04-30T00:36:36.847940920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.848588 containerd[1704]: time="2025-04-30T00:36:36.848094800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.848588 containerd[1704]: time="2025-04-30T00:36:36.848119720Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Apr 30 00:36:36.848588 containerd[1704]: time="2025-04-30T00:36:36.848145040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.848715 containerd[1704]: time="2025-04-30T00:36:36.848698680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.848782 containerd[1704]: time="2025-04-30T00:36:36.848768680Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 30 00:36:36.848930 containerd[1704]: time="2025-04-30T00:36:36.848903240Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 30 00:36:36.849021 containerd[1704]: time="2025-04-30T00:36:36.849004760Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Apr 30 00:36:36.851119 containerd[1704]: time="2025-04-30T00:36:36.849082200Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 30 00:36:36.851119 containerd[1704]: time="2025-04-30T00:36:36.849101360Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Apr 30 00:36:36.851119 containerd[1704]: time="2025-04-30T00:36:36.849113840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.851119 containerd[1704]: time="2025-04-30T00:36:36.849127040Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Apr 30 00:36:36.851119 containerd[1704]: time="2025-04-30T00:36:36.849139880Z" level=info msg="NRI interface is disabled by configuration." Apr 30 00:36:36.851119 containerd[1704]: time="2025-04-30T00:36:36.849156880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.849468560Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.849528120Z" level=info msg="Connect containerd service" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.849555560Z" level=info msg="using legacy CRI server" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.849562240Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.849651720Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.850304480Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.850480480Z" level=info msg="Start subscribing containerd event" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.850530080Z" level=info msg="Start recovering state" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.850600720Z" level=info msg="Start event monitor" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.850612320Z" level=info msg="Start snapshots syncer" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.850623960Z" level=info msg="Start cni network conf syncer for default" Apr 30 00:36:36.851348 containerd[1704]: time="2025-04-30T00:36:36.850632160Z" level=info msg="Start streaming server" Apr 30 00:36:36.852446 containerd[1704]: time="2025-04-30T00:36:36.852420040Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 30 00:36:36.853127 containerd[1704]: time="2025-04-30T00:36:36.853107840Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 30 00:36:36.853429 containerd[1704]: time="2025-04-30T00:36:36.853411720Z" level=info msg="containerd successfully booted in 0.105610s" Apr 30 00:36:36.853510 systemd[1]: Started containerd.service - containerd container runtime. Apr 30 00:36:37.192806 kubelet[1780]: E0430 00:36:37.192657 1780 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 00:36:37.195444 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 00:36:37.195730 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 00:36:37.970297 sshd_keygen[1677]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 30 00:36:37.990535 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 30 00:36:38.001342 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 30 00:36:38.008353 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Apr 30 00:36:38.015784 systemd[1]: issuegen.service: Deactivated successfully. Apr 30 00:36:38.018136 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 30 00:36:38.035582 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 30 00:36:38.044143 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Apr 30 00:36:38.061240 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 30 00:36:38.075555 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 30 00:36:38.082679 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Apr 30 00:36:38.089436 systemd[1]: Reached target getty.target - Login Prompts. Apr 30 00:36:38.096654 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 30 00:36:38.107190 systemd[1]: Startup finished in 700ms (kernel) + 12.579s (initrd) + 12.984s (userspace) = 26.264s. Apr 30 00:36:38.324737 login[1812]: pam_lastlog(login:session): file /var/log/lastlog is locked/write, retrying Apr 30 00:36:38.325017 login[1811]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:36:38.334466 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 30 00:36:38.344416 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 30 00:36:38.348891 systemd-logind[1668]: New session 1 of user core. Apr 30 00:36:38.357735 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 30 00:36:38.366453 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 30 00:36:38.369203 (systemd)[1819]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 30 00:36:38.503602 systemd[1819]: Queued start job for default target default.target. Apr 30 00:36:38.510048 systemd[1819]: Created slice app.slice - User Application Slice. Apr 30 00:36:38.510101 systemd[1819]: Reached target paths.target - Paths. Apr 30 00:36:38.510115 systemd[1819]: Reached target timers.target - Timers. Apr 30 00:36:38.511449 systemd[1819]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 30 00:36:38.526572 systemd[1819]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 30 00:36:38.526699 systemd[1819]: Reached target sockets.target - Sockets. Apr 30 00:36:38.526713 systemd[1819]: Reached target basic.target - Basic System. Apr 30 00:36:38.526771 systemd[1819]: Reached target default.target - Main User Target. Apr 30 00:36:38.526803 systemd[1819]: Startup finished in 150ms. Apr 30 00:36:38.527192 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 30 00:36:38.535312 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 30 00:36:39.326281 login[1812]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:36:39.332684 systemd-logind[1668]: New session 2 of user core. Apr 30 00:36:39.335254 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 30 00:36:39.657078 waagent[1808]: 2025-04-30T00:36:39.656892Z INFO Daemon Daemon Azure Linux Agent Version: 2.9.1.1 Apr 30 00:36:39.663717 waagent[1808]: 2025-04-30T00:36:39.663625Z INFO Daemon Daemon OS: flatcar 4081.3.3 Apr 30 00:36:39.669153 waagent[1808]: 2025-04-30T00:36:39.669082Z INFO Daemon Daemon Python: 3.11.9 Apr 30 00:36:39.674633 waagent[1808]: 2025-04-30T00:36:39.674412Z INFO Daemon Daemon Run daemon Apr 30 00:36:39.679275 waagent[1808]: 2025-04-30T00:36:39.679033Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4081.3.3' Apr 30 00:36:39.689257 waagent[1808]: 2025-04-30T00:36:39.689161Z INFO Daemon Daemon Using waagent for provisioning Apr 30 00:36:39.695258 waagent[1808]: 2025-04-30T00:36:39.695194Z INFO Daemon Daemon Activate resource disk Apr 30 00:36:39.700235 waagent[1808]: 2025-04-30T00:36:39.700164Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Apr 30 00:36:39.711070 waagent[1808]: 2025-04-30T00:36:39.710993Z INFO Daemon Daemon Found device: None Apr 30 00:36:39.715349 waagent[1808]: 2025-04-30T00:36:39.715276Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Apr 30 00:36:39.723712 waagent[1808]: 2025-04-30T00:36:39.723642Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Apr 30 00:36:39.736830 waagent[1808]: 2025-04-30T00:36:39.736745Z INFO Daemon Daemon Clean protocol and wireserver endpoint Apr 30 00:36:39.742495 waagent[1808]: 2025-04-30T00:36:39.742422Z INFO Daemon Daemon Running default provisioning handler Apr 30 00:36:39.755111 waagent[1808]: 2025-04-30T00:36:39.754848Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Apr 30 00:36:39.769472 waagent[1808]: 2025-04-30T00:36:39.769385Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Apr 30 00:36:39.778592 waagent[1808]: 2025-04-30T00:36:39.778516Z INFO Daemon Daemon cloud-init is enabled: False Apr 30 00:36:39.783546 waagent[1808]: 2025-04-30T00:36:39.783480Z INFO Daemon Daemon Copying ovf-env.xml Apr 30 00:36:39.930147 waagent[1808]: 2025-04-30T00:36:39.928351Z INFO Daemon Daemon Successfully mounted dvd Apr 30 00:36:39.958761 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Apr 30 00:36:39.962103 waagent[1808]: 2025-04-30T00:36:39.961985Z INFO Daemon Daemon Detect protocol endpoint Apr 30 00:36:39.966987 waagent[1808]: 2025-04-30T00:36:39.966905Z INFO Daemon Daemon Clean protocol and wireserver endpoint Apr 30 00:36:39.972783 waagent[1808]: 2025-04-30T00:36:39.972709Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Apr 30 00:36:39.979274 waagent[1808]: 2025-04-30T00:36:39.979198Z INFO Daemon Daemon Test for route to 168.63.129.16 Apr 30 00:36:39.984729 waagent[1808]: 2025-04-30T00:36:39.984665Z INFO Daemon Daemon Route to 168.63.129.16 exists Apr 30 00:36:39.989949 waagent[1808]: 2025-04-30T00:36:39.989885Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Apr 30 00:36:40.025863 waagent[1808]: 2025-04-30T00:36:40.025803Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Apr 30 00:36:40.032643 waagent[1808]: 2025-04-30T00:36:40.032605Z INFO Daemon Daemon Wire protocol version:2012-11-30 Apr 30 00:36:40.037964 waagent[1808]: 2025-04-30T00:36:40.037899Z INFO Daemon Daemon Server preferred version:2015-04-05 Apr 30 00:36:40.397577 waagent[1808]: 2025-04-30T00:36:40.397404Z INFO Daemon Daemon Initializing goal state during protocol detection Apr 30 00:36:40.403866 waagent[1808]: 2025-04-30T00:36:40.403783Z INFO Daemon Daemon Forcing an update of the goal state. Apr 30 00:36:40.412823 waagent[1808]: 2025-04-30T00:36:40.412766Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Apr 30 00:36:40.433260 waagent[1808]: 2025-04-30T00:36:40.433205Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.164 Apr 30 00:36:40.438852 waagent[1808]: 2025-04-30T00:36:40.438796Z INFO Daemon Apr 30 00:36:40.441669 waagent[1808]: 2025-04-30T00:36:40.441612Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: e0a945aa-7c5c-4022-bb7b-5cd4cac1888d eTag: 11787488472974045851 source: Fabric] Apr 30 00:36:40.452710 waagent[1808]: 2025-04-30T00:36:40.452654Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Apr 30 00:36:40.460039 waagent[1808]: 2025-04-30T00:36:40.459980Z INFO Daemon Apr 30 00:36:40.463556 waagent[1808]: 2025-04-30T00:36:40.463462Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Apr 30 00:36:40.477051 waagent[1808]: 2025-04-30T00:36:40.477001Z INFO Daemon Daemon Downloading artifacts profile blob Apr 30 00:36:40.648897 waagent[1808]: 2025-04-30T00:36:40.648739Z INFO Daemon Downloaded certificate {'thumbprint': 'CCECBF30C8923E75CEA946E2894DFA12E9AA3BCD', 'hasPrivateKey': True} Apr 30 00:36:40.660443 waagent[1808]: 2025-04-30T00:36:40.660356Z INFO Daemon Downloaded certificate {'thumbprint': '9BE586681EEE541F5A513FD81CEE00A351B413E1', 'hasPrivateKey': False} Apr 30 00:36:40.671510 waagent[1808]: 2025-04-30T00:36:40.671451Z INFO Daemon Fetch goal state completed Apr 30 00:36:40.722980 waagent[1808]: 2025-04-30T00:36:40.722929Z INFO Daemon Daemon Starting provisioning Apr 30 00:36:40.728172 waagent[1808]: 2025-04-30T00:36:40.728106Z INFO Daemon Daemon Handle ovf-env.xml. Apr 30 00:36:40.732795 waagent[1808]: 2025-04-30T00:36:40.732735Z INFO Daemon Daemon Set hostname [ci-4081.3.3-a-c388cdac8a] Apr 30 00:36:40.754433 waagent[1808]: 2025-04-30T00:36:40.754349Z INFO Daemon Daemon Publish hostname [ci-4081.3.3-a-c388cdac8a] Apr 30 00:36:40.761012 waagent[1808]: 2025-04-30T00:36:40.760942Z INFO Daemon Daemon Examine /proc/net/route for primary interface Apr 30 00:36:40.767765 waagent[1808]: 2025-04-30T00:36:40.767703Z INFO Daemon Daemon Primary interface is [eth0] Apr 30 00:36:40.799638 systemd-networkd[1329]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 00:36:40.799649 systemd-networkd[1329]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 00:36:40.799683 systemd-networkd[1329]: eth0: DHCP lease lost Apr 30 00:36:40.802094 waagent[1808]: 2025-04-30T00:36:40.801417Z INFO Daemon Daemon Create user account if not exists Apr 30 00:36:40.812502 waagent[1808]: 2025-04-30T00:36:40.807101Z INFO Daemon Daemon User core already exists, skip useradd Apr 30 00:36:40.807171 systemd-networkd[1329]: eth0: DHCPv6 lease lost Apr 30 00:36:40.812875 waagent[1808]: 2025-04-30T00:36:40.812797Z INFO Daemon Daemon Configure sudoer Apr 30 00:36:40.817467 waagent[1808]: 2025-04-30T00:36:40.817388Z INFO Daemon Daemon Configure sshd Apr 30 00:36:40.821900 waagent[1808]: 2025-04-30T00:36:40.821832Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Apr 30 00:36:40.834715 waagent[1808]: 2025-04-30T00:36:40.834408Z INFO Daemon Daemon Deploy ssh public key. Apr 30 00:36:40.844146 systemd-networkd[1329]: eth0: DHCPv4 address 10.200.20.12/24, gateway 10.200.20.1 acquired from 168.63.129.16 Apr 30 00:36:41.967225 waagent[1808]: 2025-04-30T00:36:41.967159Z INFO Daemon Daemon Provisioning complete Apr 30 00:36:41.984819 waagent[1808]: 2025-04-30T00:36:41.984764Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Apr 30 00:36:41.991006 waagent[1808]: 2025-04-30T00:36:41.990935Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Apr 30 00:36:42.000963 waagent[1808]: 2025-04-30T00:36:42.000893Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.9.1.1 is the most current agent Apr 30 00:36:42.149124 waagent[1874]: 2025-04-30T00:36:42.148287Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.9.1.1) Apr 30 00:36:42.149124 waagent[1874]: 2025-04-30T00:36:42.148471Z INFO ExtHandler ExtHandler OS: flatcar 4081.3.3 Apr 30 00:36:42.149124 waagent[1874]: 2025-04-30T00:36:42.148525Z INFO ExtHandler ExtHandler Python: 3.11.9 Apr 30 00:36:42.174958 waagent[1874]: 2025-04-30T00:36:42.174842Z INFO ExtHandler ExtHandler Distro: flatcar-4081.3.3; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.9; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Apr 30 00:36:42.175203 waagent[1874]: 2025-04-30T00:36:42.175158Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 30 00:36:42.175273 waagent[1874]: 2025-04-30T00:36:42.175242Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 30 00:36:42.184174 waagent[1874]: 2025-04-30T00:36:42.184033Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Apr 30 00:36:42.191236 waagent[1874]: 2025-04-30T00:36:42.191180Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.164 Apr 30 00:36:42.191820 waagent[1874]: 2025-04-30T00:36:42.191772Z INFO ExtHandler Apr 30 00:36:42.191900 waagent[1874]: 2025-04-30T00:36:42.191868Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 19fd6aba-394e-41f3-ad33-8959017b86d3 eTag: 11787488472974045851 source: Fabric] Apr 30 00:36:42.192236 waagent[1874]: 2025-04-30T00:36:42.192191Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Apr 30 00:36:42.192841 waagent[1874]: 2025-04-30T00:36:42.192792Z INFO ExtHandler Apr 30 00:36:42.192913 waagent[1874]: 2025-04-30T00:36:42.192882Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Apr 30 00:36:42.197320 waagent[1874]: 2025-04-30T00:36:42.197274Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Apr 30 00:36:42.288536 waagent[1874]: 2025-04-30T00:36:42.288372Z INFO ExtHandler Downloaded certificate {'thumbprint': 'CCECBF30C8923E75CEA946E2894DFA12E9AA3BCD', 'hasPrivateKey': True} Apr 30 00:36:42.288932 waagent[1874]: 2025-04-30T00:36:42.288883Z INFO ExtHandler Downloaded certificate {'thumbprint': '9BE586681EEE541F5A513FD81CEE00A351B413E1', 'hasPrivateKey': False} Apr 30 00:36:42.289431 waagent[1874]: 2025-04-30T00:36:42.289381Z INFO ExtHandler Fetch goal state completed Apr 30 00:36:42.306437 waagent[1874]: 2025-04-30T00:36:42.306362Z INFO ExtHandler ExtHandler WALinuxAgent-2.9.1.1 running as process 1874 Apr 30 00:36:42.306623 waagent[1874]: 2025-04-30T00:36:42.306581Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Apr 30 00:36:42.308445 waagent[1874]: 2025-04-30T00:36:42.308384Z INFO ExtHandler ExtHandler Cgroup monitoring is not supported on ['flatcar', '4081.3.3', '', 'Flatcar Container Linux by Kinvolk'] Apr 30 00:36:42.308862 waagent[1874]: 2025-04-30T00:36:42.308820Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Apr 30 00:36:42.340886 waagent[1874]: 2025-04-30T00:36:42.340834Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Apr 30 00:36:42.341113 waagent[1874]: 2025-04-30T00:36:42.341048Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Apr 30 00:36:42.348099 waagent[1874]: 2025-04-30T00:36:42.347763Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Apr 30 00:36:42.355709 systemd[1]: Reloading requested from client PID 1889 ('systemctl') (unit waagent.service)... Apr 30 00:36:42.355997 systemd[1]: Reloading... Apr 30 00:36:42.444105 zram_generator::config[1926]: No configuration found. Apr 30 00:36:42.564197 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 00:36:42.644819 systemd[1]: Reloading finished in 288 ms. Apr 30 00:36:42.673103 waagent[1874]: 2025-04-30T00:36:42.670509Z INFO ExtHandler ExtHandler Executing systemctl daemon-reload for setting up waagent-network-setup.service Apr 30 00:36:42.677479 systemd[1]: Reloading requested from client PID 1977 ('systemctl') (unit waagent.service)... Apr 30 00:36:42.677496 systemd[1]: Reloading... Apr 30 00:36:42.741217 zram_generator::config[2011]: No configuration found. Apr 30 00:36:42.875897 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 00:36:42.957466 systemd[1]: Reloading finished in 279 ms. Apr 30 00:36:42.984953 waagent[1874]: 2025-04-30T00:36:42.981594Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Apr 30 00:36:42.984953 waagent[1874]: 2025-04-30T00:36:42.981790Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Apr 30 00:36:43.322677 waagent[1874]: 2025-04-30T00:36:43.321243Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Apr 30 00:36:43.322677 waagent[1874]: 2025-04-30T00:36:43.321863Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: configuration enabled [True], cgroups enabled [False], python supported: [True] Apr 30 00:36:43.323014 waagent[1874]: 2025-04-30T00:36:43.322903Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 30 00:36:43.323014 waagent[1874]: 2025-04-30T00:36:43.322993Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 30 00:36:43.323269 waagent[1874]: 2025-04-30T00:36:43.323222Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Apr 30 00:36:43.323393 waagent[1874]: 2025-04-30T00:36:43.323330Z INFO ExtHandler ExtHandler Starting env monitor service. Apr 30 00:36:43.323905 waagent[1874]: 2025-04-30T00:36:43.323853Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Apr 30 00:36:43.323905 waagent[1874]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Apr 30 00:36:43.323905 waagent[1874]: eth0 00000000 0114C80A 0003 0 0 1024 00000000 0 0 0 Apr 30 00:36:43.323905 waagent[1874]: eth0 0014C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Apr 30 00:36:43.323905 waagent[1874]: eth0 0114C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Apr 30 00:36:43.323905 waagent[1874]: eth0 10813FA8 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Apr 30 00:36:43.323905 waagent[1874]: eth0 FEA9FEA9 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Apr 30 00:36:43.324163 waagent[1874]: 2025-04-30T00:36:43.324103Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Apr 30 00:36:43.324721 waagent[1874]: 2025-04-30T00:36:43.324666Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Apr 30 00:36:43.324858 waagent[1874]: 2025-04-30T00:36:43.324818Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 30 00:36:43.324929 waagent[1874]: 2025-04-30T00:36:43.324899Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 30 00:36:43.325096 waagent[1874]: 2025-04-30T00:36:43.325040Z INFO EnvHandler ExtHandler Configure routes Apr 30 00:36:43.325170 waagent[1874]: 2025-04-30T00:36:43.325138Z INFO EnvHandler ExtHandler Gateway:None Apr 30 00:36:43.325222 waagent[1874]: 2025-04-30T00:36:43.325195Z INFO EnvHandler ExtHandler Routes:None Apr 30 00:36:43.325648 waagent[1874]: 2025-04-30T00:36:43.325587Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Apr 30 00:36:43.326813 waagent[1874]: 2025-04-30T00:36:43.326244Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Apr 30 00:36:43.326813 waagent[1874]: 2025-04-30T00:36:43.326367Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Apr 30 00:36:43.326992 waagent[1874]: 2025-04-30T00:36:43.326940Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Apr 30 00:36:43.334092 waagent[1874]: 2025-04-30T00:36:43.333991Z INFO ExtHandler ExtHandler Apr 30 00:36:43.334689 waagent[1874]: 2025-04-30T00:36:43.334630Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: a8036fcf-fe0f-4122-8e35-2b86058df46e correlation eaced721-35a4-462b-b979-002189acd3fa created: 2025-04-30T00:35:24.144600Z] Apr 30 00:36:43.335841 waagent[1874]: 2025-04-30T00:36:43.335768Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Apr 30 00:36:43.337028 waagent[1874]: 2025-04-30T00:36:43.336967Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 2 ms] Apr 30 00:36:43.365521 waagent[1874]: 2025-04-30T00:36:43.365427Z INFO MonitorHandler ExtHandler Network interfaces: Apr 30 00:36:43.365521 waagent[1874]: Executing ['ip', '-a', '-o', 'link']: Apr 30 00:36:43.365521 waagent[1874]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Apr 30 00:36:43.365521 waagent[1874]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 00:22:48:b7:44:a6 brd ff:ff:ff:ff:ff:ff Apr 30 00:36:43.365521 waagent[1874]: 3: enP31290s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 00:22:48:b7:44:a6 brd ff:ff:ff:ff:ff:ff\ altname enP31290p0s2 Apr 30 00:36:43.365521 waagent[1874]: Executing ['ip', '-4', '-a', '-o', 'address']: Apr 30 00:36:43.365521 waagent[1874]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Apr 30 00:36:43.365521 waagent[1874]: 2: eth0 inet 10.200.20.12/24 metric 1024 brd 10.200.20.255 scope global eth0\ valid_lft forever preferred_lft forever Apr 30 00:36:43.365521 waagent[1874]: Executing ['ip', '-6', '-a', '-o', 'address']: Apr 30 00:36:43.365521 waagent[1874]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Apr 30 00:36:43.365521 waagent[1874]: 2: eth0 inet6 fe80::222:48ff:feb7:44a6/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Apr 30 00:36:43.365521 waagent[1874]: 3: enP31290s1 inet6 fe80::222:48ff:feb7:44a6/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Apr 30 00:36:43.398129 waagent[1874]: 2025-04-30T00:36:43.397999Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.9.1.1 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 90A205E9-5C27-44A6-B5C8-9E4F232D2BB6;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 0] Apr 30 00:36:43.431767 waagent[1874]: 2025-04-30T00:36:43.431658Z INFO EnvHandler ExtHandler Successfully added Azure fabric firewall rules. Current Firewall rules: Apr 30 00:36:43.431767 waagent[1874]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Apr 30 00:36:43.431767 waagent[1874]: pkts bytes target prot opt in out source destination Apr 30 00:36:43.431767 waagent[1874]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Apr 30 00:36:43.431767 waagent[1874]: pkts bytes target prot opt in out source destination Apr 30 00:36:43.431767 waagent[1874]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Apr 30 00:36:43.431767 waagent[1874]: pkts bytes target prot opt in out source destination Apr 30 00:36:43.431767 waagent[1874]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Apr 30 00:36:43.431767 waagent[1874]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Apr 30 00:36:43.431767 waagent[1874]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Apr 30 00:36:43.435486 waagent[1874]: 2025-04-30T00:36:43.435398Z INFO EnvHandler ExtHandler Current Firewall rules: Apr 30 00:36:43.435486 waagent[1874]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Apr 30 00:36:43.435486 waagent[1874]: pkts bytes target prot opt in out source destination Apr 30 00:36:43.435486 waagent[1874]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Apr 30 00:36:43.435486 waagent[1874]: pkts bytes target prot opt in out source destination Apr 30 00:36:43.435486 waagent[1874]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Apr 30 00:36:43.435486 waagent[1874]: pkts bytes target prot opt in out source destination Apr 30 00:36:43.435486 waagent[1874]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Apr 30 00:36:43.435486 waagent[1874]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Apr 30 00:36:43.435486 waagent[1874]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Apr 30 00:36:43.435775 waagent[1874]: 2025-04-30T00:36:43.435731Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Apr 30 00:36:47.363175 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 30 00:36:47.373333 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:36:47.481931 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:36:47.493387 (kubelet)[2107]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 00:36:47.547582 kubelet[2107]: E0430 00:36:47.547530 2107 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 00:36:47.550986 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 00:36:47.551317 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 00:36:57.613303 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Apr 30 00:36:57.619297 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:36:57.990530 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:36:57.995544 (kubelet)[2123]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 00:36:58.039406 kubelet[2123]: E0430 00:36:58.039284 2123 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 00:36:58.041701 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 00:36:58.041855 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 00:36:59.367717 chronyd[1652]: Selected source PHC0 Apr 30 00:37:08.113347 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Apr 30 00:37:08.122306 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:37:08.220312 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:37:08.222993 (kubelet)[2138]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 00:37:08.263148 kubelet[2138]: E0430 00:37:08.263029 2138 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 00:37:08.264851 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 00:37:08.264980 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 00:37:10.949520 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 30 00:37:10.950784 systemd[1]: Started sshd@0-10.200.20.12:22-10.200.16.10:34536.service - OpenSSH per-connection server daemon (10.200.16.10:34536). Apr 30 00:37:11.469668 sshd[2145]: Accepted publickey for core from 10.200.16.10 port 34536 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:37:11.471151 sshd[2145]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:37:11.475192 systemd-logind[1668]: New session 3 of user core. Apr 30 00:37:11.486240 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 30 00:37:11.856369 systemd[1]: Started sshd@1-10.200.20.12:22-10.200.16.10:34544.service - OpenSSH per-connection server daemon (10.200.16.10:34544). Apr 30 00:37:12.304966 sshd[2150]: Accepted publickey for core from 10.200.16.10 port 34544 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:37:12.306386 sshd[2150]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:37:12.311979 systemd-logind[1668]: New session 4 of user core. Apr 30 00:37:12.317269 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 30 00:37:12.630973 sshd[2150]: pam_unix(sshd:session): session closed for user core Apr 30 00:37:12.634602 systemd[1]: sshd@1-10.200.20.12:22-10.200.16.10:34544.service: Deactivated successfully. Apr 30 00:37:12.636423 systemd[1]: session-4.scope: Deactivated successfully. Apr 30 00:37:12.638152 systemd-logind[1668]: Session 4 logged out. Waiting for processes to exit. Apr 30 00:37:12.640513 systemd-logind[1668]: Removed session 4. Apr 30 00:37:12.706694 systemd[1]: Started sshd@2-10.200.20.12:22-10.200.16.10:34560.service - OpenSSH per-connection server daemon (10.200.16.10:34560). Apr 30 00:37:13.123248 sshd[2157]: Accepted publickey for core from 10.200.16.10 port 34560 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:37:13.126579 sshd[2157]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:37:13.130945 systemd-logind[1668]: New session 5 of user core. Apr 30 00:37:13.138271 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 30 00:37:13.423638 sshd[2157]: pam_unix(sshd:session): session closed for user core Apr 30 00:37:13.427371 systemd-logind[1668]: Session 5 logged out. Waiting for processes to exit. Apr 30 00:37:13.429170 systemd[1]: sshd@2-10.200.20.12:22-10.200.16.10:34560.service: Deactivated successfully. Apr 30 00:37:13.431257 systemd[1]: session-5.scope: Deactivated successfully. Apr 30 00:37:13.432141 systemd-logind[1668]: Removed session 5. Apr 30 00:37:13.531354 systemd[1]: Started sshd@3-10.200.20.12:22-10.200.16.10:34576.service - OpenSSH per-connection server daemon (10.200.16.10:34576). Apr 30 00:37:13.974376 sshd[2164]: Accepted publickey for core from 10.200.16.10 port 34576 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:37:13.975805 sshd[2164]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:37:13.981144 systemd-logind[1668]: New session 6 of user core. Apr 30 00:37:13.986295 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 30 00:37:14.300017 sshd[2164]: pam_unix(sshd:session): session closed for user core Apr 30 00:37:14.303982 systemd[1]: sshd@3-10.200.20.12:22-10.200.16.10:34576.service: Deactivated successfully. Apr 30 00:37:14.305770 systemd[1]: session-6.scope: Deactivated successfully. Apr 30 00:37:14.307593 systemd-logind[1668]: Session 6 logged out. Waiting for processes to exit. Apr 30 00:37:14.308676 systemd-logind[1668]: Removed session 6. Apr 30 00:37:14.388351 systemd[1]: Started sshd@4-10.200.20.12:22-10.200.16.10:34586.service - OpenSSH per-connection server daemon (10.200.16.10:34586). Apr 30 00:37:14.830243 sshd[2171]: Accepted publickey for core from 10.200.16.10 port 34586 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:37:14.831590 sshd[2171]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:37:14.835596 systemd-logind[1668]: New session 7 of user core. Apr 30 00:37:14.846285 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 30 00:37:15.243718 sudo[2174]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 30 00:37:15.244370 sudo[2174]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 00:37:15.277536 sudo[2174]: pam_unix(sudo:session): session closed for user root Apr 30 00:37:15.347925 sshd[2171]: pam_unix(sshd:session): session closed for user core Apr 30 00:37:15.351896 systemd[1]: sshd@4-10.200.20.12:22-10.200.16.10:34586.service: Deactivated successfully. Apr 30 00:37:15.353598 systemd[1]: session-7.scope: Deactivated successfully. Apr 30 00:37:15.354324 systemd-logind[1668]: Session 7 logged out. Waiting for processes to exit. Apr 30 00:37:15.355647 systemd-logind[1668]: Removed session 7. Apr 30 00:37:15.435021 systemd[1]: Started sshd@5-10.200.20.12:22-10.200.16.10:34588.service - OpenSSH per-connection server daemon (10.200.16.10:34588). Apr 30 00:37:15.885217 sshd[2179]: Accepted publickey for core from 10.200.16.10 port 34588 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:37:15.886590 sshd[2179]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:37:15.891459 systemd-logind[1668]: New session 8 of user core. Apr 30 00:37:15.897281 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 30 00:37:16.141168 sudo[2183]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 30 00:37:16.141893 sudo[2183]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 00:37:16.145352 sudo[2183]: pam_unix(sudo:session): session closed for user root Apr 30 00:37:16.150698 sudo[2182]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Apr 30 00:37:16.150982 sudo[2182]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 00:37:16.171560 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Apr 30 00:37:16.172834 auditctl[2186]: No rules Apr 30 00:37:16.173224 systemd[1]: audit-rules.service: Deactivated successfully. Apr 30 00:37:16.173412 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Apr 30 00:37:16.176851 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 30 00:37:16.203922 augenrules[2204]: No rules Apr 30 00:37:16.205781 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 30 00:37:16.207372 sudo[2182]: pam_unix(sudo:session): session closed for user root Apr 30 00:37:16.278344 sshd[2179]: pam_unix(sshd:session): session closed for user core Apr 30 00:37:16.282455 systemd-logind[1668]: Session 8 logged out. Waiting for processes to exit. Apr 30 00:37:16.283790 systemd[1]: sshd@5-10.200.20.12:22-10.200.16.10:34588.service: Deactivated successfully. Apr 30 00:37:16.286795 systemd[1]: session-8.scope: Deactivated successfully. Apr 30 00:37:16.287922 systemd-logind[1668]: Removed session 8. Apr 30 00:37:16.353181 systemd[1]: Started sshd@6-10.200.20.12:22-10.200.16.10:34598.service - OpenSSH per-connection server daemon (10.200.16.10:34598). Apr 30 00:37:16.765675 sshd[2212]: Accepted publickey for core from 10.200.16.10 port 34598 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:37:16.767108 sshd[2212]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:37:16.770907 systemd-logind[1668]: New session 9 of user core. Apr 30 00:37:16.781247 systemd[1]: Started session-9.scope - Session 9 of User core. Apr 30 00:37:17.003410 sudo[2215]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 30 00:37:17.003690 sudo[2215]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 00:37:18.363108 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Apr 30 00:37:18.372291 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:37:19.057292 kernel: hv_balloon: Max. dynamic memory size: 4096 MB Apr 30 00:37:20.724333 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 30 00:37:20.724530 (dockerd)[2233]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 30 00:37:20.848017 update_engine[1671]: I20250430 00:37:20.847922 1671 update_attempter.cc:509] Updating boot flags... Apr 30 00:37:21.206148 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (2245) Apr 30 00:37:21.456311 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (2238) Apr 30 00:37:21.756129 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (2238) Apr 30 00:37:22.954088 dockerd[2233]: time="2025-04-30T00:37:22.953884791Z" level=info msg="Starting up" Apr 30 00:37:24.461305 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:37:24.466336 (kubelet)[2351]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 00:37:24.503794 kubelet[2351]: E0430 00:37:24.503732 2351 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 00:37:24.506333 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 00:37:24.506487 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 00:37:26.512712 systemd[1]: var-lib-docker-metacopy\x2dcheck1386565148-merged.mount: Deactivated successfully. Apr 30 00:37:26.616461 dockerd[2233]: time="2025-04-30T00:37:26.616062214Z" level=info msg="Loading containers: start." Apr 30 00:37:26.857100 kernel: Initializing XFRM netlink socket Apr 30 00:37:27.051691 systemd-networkd[1329]: docker0: Link UP Apr 30 00:37:27.088004 dockerd[2233]: time="2025-04-30T00:37:27.087377752Z" level=info msg="Loading containers: done." Apr 30 00:37:27.098597 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3598625595-merged.mount: Deactivated successfully. Apr 30 00:37:27.668612 dockerd[2233]: time="2025-04-30T00:37:27.668562091Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 30 00:37:27.669048 dockerd[2233]: time="2025-04-30T00:37:27.668704131Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Apr 30 00:37:27.669048 dockerd[2233]: time="2025-04-30T00:37:27.668858971Z" level=info msg="Daemon has completed initialization" Apr 30 00:37:27.867316 dockerd[2233]: time="2025-04-30T00:37:27.867227166Z" level=info msg="API listen on /run/docker.sock" Apr 30 00:37:27.867649 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 30 00:37:28.684292 containerd[1704]: time="2025-04-30T00:37:28.684245634Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.8\"" Apr 30 00:37:30.521205 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount315655632.mount: Deactivated successfully. Apr 30 00:37:32.246116 containerd[1704]: time="2025-04-30T00:37:32.245167771Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:32.248812 containerd[1704]: time="2025-04-30T00:37:32.248581134Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.8: active requests=0, bytes read=25554608" Apr 30 00:37:32.253438 containerd[1704]: time="2025-04-30T00:37:32.253370738Z" level=info msg="ImageCreate event name:\"sha256:ef8fb1ea7c9599dbedea6f9d5589975ebc5bf4ec72f6be6acaaec59a723a09b3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:32.259753 containerd[1704]: time="2025-04-30T00:37:32.259674063Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:30090db6a7d53799163ce82dae9e8ddb645fd47db93f2ec9da0cc787fd825625\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:32.261051 containerd[1704]: time="2025-04-30T00:37:32.260844144Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.8\" with image id \"sha256:ef8fb1ea7c9599dbedea6f9d5589975ebc5bf4ec72f6be6acaaec59a723a09b3\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.8\", repo digest \"registry.k8s.io/kube-apiserver@sha256:30090db6a7d53799163ce82dae9e8ddb645fd47db93f2ec9da0cc787fd825625\", size \"25551408\" in 3.57654867s" Apr 30 00:37:32.261051 containerd[1704]: time="2025-04-30T00:37:32.260891064Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.8\" returns image reference \"sha256:ef8fb1ea7c9599dbedea6f9d5589975ebc5bf4ec72f6be6acaaec59a723a09b3\"" Apr 30 00:37:32.261835 containerd[1704]: time="2025-04-30T00:37:32.261807705Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.8\"" Apr 30 00:37:33.938868 containerd[1704]: time="2025-04-30T00:37:33.938801894Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:33.947250 containerd[1704]: time="2025-04-30T00:37:33.946978261Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.8: active requests=0, bytes read=22458978" Apr 30 00:37:33.958107 containerd[1704]: time="2025-04-30T00:37:33.957980310Z" level=info msg="ImageCreate event name:\"sha256:ea6e6085feca75547d0422ab0536fe0d18c9ff5831de7a9d6a707c968027bb6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:33.967147 containerd[1704]: time="2025-04-30T00:37:33.967058798Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:29eaddc64792a689df48506e78bbc641d063ac8bb92d2e66ae2ad05977420747\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:33.968543 containerd[1704]: time="2025-04-30T00:37:33.968411759Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.8\" with image id \"sha256:ea6e6085feca75547d0422ab0536fe0d18c9ff5831de7a9d6a707c968027bb6a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.8\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:29eaddc64792a689df48506e78bbc641d063ac8bb92d2e66ae2ad05977420747\", size \"23900539\" in 1.706561934s" Apr 30 00:37:33.968543 containerd[1704]: time="2025-04-30T00:37:33.968450519Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.8\" returns image reference \"sha256:ea6e6085feca75547d0422ab0536fe0d18c9ff5831de7a9d6a707c968027bb6a\"" Apr 30 00:37:33.969197 containerd[1704]: time="2025-04-30T00:37:33.969175360Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.8\"" Apr 30 00:37:34.613203 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Apr 30 00:37:34.621284 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:37:34.733190 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:37:34.738042 (kubelet)[2542]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 00:37:34.782576 kubelet[2542]: E0430 00:37:34.782509 2542 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 00:37:34.785301 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 00:37:34.785596 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 00:37:35.621117 containerd[1704]: time="2025-04-30T00:37:35.620115647Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:35.628100 containerd[1704]: time="2025-04-30T00:37:35.627979053Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.8: active requests=0, bytes read=17125813" Apr 30 00:37:35.637528 containerd[1704]: time="2025-04-30T00:37:35.637466461Z" level=info msg="ImageCreate event name:\"sha256:1d2db6ef0dd2f3e08bdfcd46afde7b755b05192841f563d8df54b807daaa7d8d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:35.645688 containerd[1704]: time="2025-04-30T00:37:35.645606708Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:22994a2632e81059720480b9f6bdeb133b08d58492d0b36dfd6e9768b159b22a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:35.647237 containerd[1704]: time="2025-04-30T00:37:35.646840549Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.8\" with image id \"sha256:1d2db6ef0dd2f3e08bdfcd46afde7b755b05192841f563d8df54b807daaa7d8d\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.8\", repo digest \"registry.k8s.io/kube-scheduler@sha256:22994a2632e81059720480b9f6bdeb133b08d58492d0b36dfd6e9768b159b22a\", size \"18567392\" in 1.677559069s" Apr 30 00:37:35.647237 containerd[1704]: time="2025-04-30T00:37:35.646880749Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.8\" returns image reference \"sha256:1d2db6ef0dd2f3e08bdfcd46afde7b755b05192841f563d8df54b807daaa7d8d\"" Apr 30 00:37:35.647456 containerd[1704]: time="2025-04-30T00:37:35.647424190Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.8\"" Apr 30 00:37:37.039422 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3341898919.mount: Deactivated successfully. Apr 30 00:37:37.395992 containerd[1704]: time="2025-04-30T00:37:37.395246279Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:37.402147 containerd[1704]: time="2025-04-30T00:37:37.402102885Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.8: active requests=0, bytes read=26871917" Apr 30 00:37:37.407239 containerd[1704]: time="2025-04-30T00:37:37.407174929Z" level=info msg="ImageCreate event name:\"sha256:c5361ece77e80334cd5fb082c0b678cb3244f5834ecacea1719ae6b38b465581\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:37.413687 containerd[1704]: time="2025-04-30T00:37:37.413597375Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:dd0c9a37670f209947b1ed880f06a2e93e1d41da78c037f52f94b13858769838\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:37.417105 containerd[1704]: time="2025-04-30T00:37:37.415625536Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.8\" with image id \"sha256:c5361ece77e80334cd5fb082c0b678cb3244f5834ecacea1719ae6b38b465581\", repo tag \"registry.k8s.io/kube-proxy:v1.31.8\", repo digest \"registry.k8s.io/kube-proxy@sha256:dd0c9a37670f209947b1ed880f06a2e93e1d41da78c037f52f94b13858769838\", size \"26870936\" in 1.768161866s" Apr 30 00:37:37.417105 containerd[1704]: time="2025-04-30T00:37:37.415683777Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.8\" returns image reference \"sha256:c5361ece77e80334cd5fb082c0b678cb3244f5834ecacea1719ae6b38b465581\"" Apr 30 00:37:37.418738 containerd[1704]: time="2025-04-30T00:37:37.418695699Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Apr 30 00:37:38.161911 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3651441383.mount: Deactivated successfully. Apr 30 00:37:39.807182 containerd[1704]: time="2025-04-30T00:37:39.807127162Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:39.812213 containerd[1704]: time="2025-04-30T00:37:39.812166126Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485381" Apr 30 00:37:39.823520 containerd[1704]: time="2025-04-30T00:37:39.823460216Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:39.832104 containerd[1704]: time="2025-04-30T00:37:39.830185662Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:39.833745 containerd[1704]: time="2025-04-30T00:37:39.833696625Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 2.414814246s" Apr 30 00:37:39.833886 containerd[1704]: time="2025-04-30T00:37:39.833860625Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" Apr 30 00:37:39.836110 containerd[1704]: time="2025-04-30T00:37:39.836033907Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Apr 30 00:37:40.588194 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount910850252.mount: Deactivated successfully. Apr 30 00:37:40.627816 containerd[1704]: time="2025-04-30T00:37:40.627756561Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:40.632624 containerd[1704]: time="2025-04-30T00:37:40.632576685Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Apr 30 00:37:40.640329 containerd[1704]: time="2025-04-30T00:37:40.640269532Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:40.647272 containerd[1704]: time="2025-04-30T00:37:40.647211418Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:40.648322 containerd[1704]: time="2025-04-30T00:37:40.647946659Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 811.800312ms" Apr 30 00:37:40.648322 containerd[1704]: time="2025-04-30T00:37:40.647987859Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Apr 30 00:37:40.648865 containerd[1704]: time="2025-04-30T00:37:40.648610499Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Apr 30 00:37:41.598538 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3337583772.mount: Deactivated successfully. Apr 30 00:37:44.863167 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Apr 30 00:37:44.871634 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:37:45.360284 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:37:45.365226 (kubelet)[2669]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 00:37:45.413224 kubelet[2669]: E0430 00:37:45.413150 2669 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 00:37:45.415999 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 00:37:45.416252 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 00:37:46.085877 containerd[1704]: time="2025-04-30T00:37:46.084616783Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:46.087510 containerd[1704]: time="2025-04-30T00:37:46.087467786Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=66406465" Apr 30 00:37:46.091504 containerd[1704]: time="2025-04-30T00:37:46.091444629Z" level=info msg="ImageCreate event name:\"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:46.107121 containerd[1704]: time="2025-04-30T00:37:46.106193682Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:37:46.107273 containerd[1704]: time="2025-04-30T00:37:46.107143923Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"66535646\" in 5.458495984s" Apr 30 00:37:46.107273 containerd[1704]: time="2025-04-30T00:37:46.107186563Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Apr 30 00:37:52.593486 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:37:52.600347 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:37:52.631924 systemd[1]: Reloading requested from client PID 2704 ('systemctl') (unit session-9.scope)... Apr 30 00:37:52.632115 systemd[1]: Reloading... Apr 30 00:37:52.748221 zram_generator::config[2747]: No configuration found. Apr 30 00:37:52.857673 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 00:37:52.938505 systemd[1]: Reloading finished in 306 ms. Apr 30 00:37:52.988564 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Apr 30 00:37:52.988665 systemd[1]: kubelet.service: Failed with result 'signal'. Apr 30 00:37:52.988980 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:37:52.995589 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:37:53.111598 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:37:53.117713 (kubelet)[2811]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 30 00:37:53.156865 kubelet[2811]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 00:37:53.156865 kubelet[2811]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Apr 30 00:37:53.156865 kubelet[2811]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 00:37:53.157257 kubelet[2811]: I0430 00:37:53.156933 2811 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 30 00:37:54.622174 kubelet[2811]: I0430 00:37:54.622123 2811 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Apr 30 00:37:54.622174 kubelet[2811]: I0430 00:37:54.622164 2811 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 30 00:37:54.622540 kubelet[2811]: I0430 00:37:54.622405 2811 server.go:929] "Client rotation is on, will bootstrap in background" Apr 30 00:37:54.639620 kubelet[2811]: E0430 00:37:54.639580 2811 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.200.20.12:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:54.643695 kubelet[2811]: I0430 00:37:54.643643 2811 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 30 00:37:54.650843 kubelet[2811]: E0430 00:37:54.650684 2811 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 30 00:37:54.650843 kubelet[2811]: I0430 00:37:54.650718 2811 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 30 00:37:54.654802 kubelet[2811]: I0430 00:37:54.654680 2811 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 30 00:37:54.654802 kubelet[2811]: I0430 00:37:54.654807 2811 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Apr 30 00:37:54.654967 kubelet[2811]: I0430 00:37:54.654917 2811 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 30 00:37:54.655184 kubelet[2811]: I0430 00:37:54.654944 2811 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.3-a-c388cdac8a","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 30 00:37:54.655292 kubelet[2811]: I0430 00:37:54.655193 2811 topology_manager.go:138] "Creating topology manager with none policy" Apr 30 00:37:54.655292 kubelet[2811]: I0430 00:37:54.655204 2811 container_manager_linux.go:300] "Creating device plugin manager" Apr 30 00:37:54.655361 kubelet[2811]: I0430 00:37:54.655339 2811 state_mem.go:36] "Initialized new in-memory state store" Apr 30 00:37:54.657090 kubelet[2811]: I0430 00:37:54.657031 2811 kubelet.go:408] "Attempting to sync node with API server" Apr 30 00:37:54.657090 kubelet[2811]: I0430 00:37:54.657096 2811 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 30 00:37:54.657194 kubelet[2811]: I0430 00:37:54.657132 2811 kubelet.go:314] "Adding apiserver pod source" Apr 30 00:37:54.657194 kubelet[2811]: I0430 00:37:54.657152 2811 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 30 00:37:54.659209 kubelet[2811]: W0430 00:37:54.659150 2811 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.20.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.3-a-c388cdac8a&limit=500&resourceVersion=0": dial tcp 10.200.20.12:6443: connect: connection refused Apr 30 00:37:54.659351 kubelet[2811]: E0430 00:37:54.659330 2811 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.200.20.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.3-a-c388cdac8a&limit=500&resourceVersion=0\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:54.659833 kubelet[2811]: W0430 00:37:54.659789 2811 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.20.12:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.200.20.12:6443: connect: connection refused Apr 30 00:37:54.660002 kubelet[2811]: E0430 00:37:54.659981 2811 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.200.20.12:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:54.660513 kubelet[2811]: I0430 00:37:54.660493 2811 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 30 00:37:54.662341 kubelet[2811]: I0430 00:37:54.662266 2811 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Apr 30 00:37:54.663267 kubelet[2811]: W0430 00:37:54.662761 2811 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 30 00:37:54.664197 kubelet[2811]: I0430 00:37:54.664162 2811 server.go:1269] "Started kubelet" Apr 30 00:37:54.664885 kubelet[2811]: I0430 00:37:54.664815 2811 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Apr 30 00:37:54.665777 kubelet[2811]: I0430 00:37:54.665752 2811 server.go:460] "Adding debug handlers to kubelet server" Apr 30 00:37:54.666975 kubelet[2811]: I0430 00:37:54.666916 2811 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 30 00:37:54.667363 kubelet[2811]: I0430 00:37:54.667342 2811 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 30 00:37:54.669630 kubelet[2811]: I0430 00:37:54.668185 2811 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 30 00:37:54.669630 kubelet[2811]: E0430 00:37:54.667611 2811 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.20.12:6443/api/v1/namespaces/default/events\": dial tcp 10.200.20.12:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.3-a-c388cdac8a.183af19fe89c464a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.3-a-c388cdac8a,UID:ci-4081.3.3-a-c388cdac8a,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.3-a-c388cdac8a,},FirstTimestamp:2025-04-30 00:37:54.664138314 +0000 UTC m=+1.542761360,LastTimestamp:2025-04-30 00:37:54.664138314 +0000 UTC m=+1.542761360,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.3-a-c388cdac8a,}" Apr 30 00:37:54.669630 kubelet[2811]: I0430 00:37:54.669403 2811 volume_manager.go:289] "Starting Kubelet Volume Manager" Apr 30 00:37:54.669977 kubelet[2811]: E0430 00:37:54.669935 2811 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.3-a-c388cdac8a\" not found" Apr 30 00:37:54.670503 kubelet[2811]: I0430 00:37:54.670470 2811 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 30 00:37:54.670938 kubelet[2811]: I0430 00:37:54.670890 2811 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 30 00:37:54.671046 kubelet[2811]: I0430 00:37:54.671028 2811 reconciler.go:26] "Reconciler: start to sync state" Apr 30 00:37:54.672288 kubelet[2811]: W0430 00:37:54.672221 2811 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.20.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.12:6443: connect: connection refused Apr 30 00:37:54.672367 kubelet[2811]: E0430 00:37:54.672309 2811 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.200.20.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:54.674144 kubelet[2811]: E0430 00:37:54.674055 2811 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.3-a-c388cdac8a?timeout=10s\": dial tcp 10.200.20.12:6443: connect: connection refused" interval="200ms" Apr 30 00:37:54.679745 kubelet[2811]: I0430 00:37:54.679715 2811 factory.go:221] Registration of the containerd container factory successfully Apr 30 00:37:54.679889 kubelet[2811]: I0430 00:37:54.679879 2811 factory.go:221] Registration of the systemd container factory successfully Apr 30 00:37:54.680058 kubelet[2811]: I0430 00:37:54.680038 2811 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 30 00:37:54.713650 kubelet[2811]: I0430 00:37:54.713601 2811 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Apr 30 00:37:54.714984 kubelet[2811]: I0430 00:37:54.714933 2811 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Apr 30 00:37:54.714984 kubelet[2811]: I0430 00:37:54.714967 2811 status_manager.go:217] "Starting to sync pod status with apiserver" Apr 30 00:37:54.714984 kubelet[2811]: I0430 00:37:54.714987 2811 kubelet.go:2321] "Starting kubelet main sync loop" Apr 30 00:37:54.715157 kubelet[2811]: E0430 00:37:54.715037 2811 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 30 00:37:54.719718 kubelet[2811]: W0430 00:37:54.719670 2811 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.20.12:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.12:6443: connect: connection refused Apr 30 00:37:54.719860 kubelet[2811]: E0430 00:37:54.719736 2811 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.200.20.12:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:54.720247 kubelet[2811]: I0430 00:37:54.720225 2811 cpu_manager.go:214] "Starting CPU manager" policy="none" Apr 30 00:37:54.720247 kubelet[2811]: I0430 00:37:54.720243 2811 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Apr 30 00:37:54.720334 kubelet[2811]: I0430 00:37:54.720266 2811 state_mem.go:36] "Initialized new in-memory state store" Apr 30 00:37:54.729630 kubelet[2811]: I0430 00:37:54.729599 2811 policy_none.go:49] "None policy: Start" Apr 30 00:37:54.730549 kubelet[2811]: I0430 00:37:54.730516 2811 memory_manager.go:170] "Starting memorymanager" policy="None" Apr 30 00:37:54.730648 kubelet[2811]: I0430 00:37:54.730559 2811 state_mem.go:35] "Initializing new in-memory state store" Apr 30 00:37:54.743790 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Apr 30 00:37:54.756935 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Apr 30 00:37:54.759965 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Apr 30 00:37:54.763950 kubelet[2811]: I0430 00:37:54.763910 2811 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 30 00:37:54.764169 kubelet[2811]: I0430 00:37:54.764147 2811 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 30 00:37:54.764217 kubelet[2811]: I0430 00:37:54.764166 2811 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 30 00:37:54.765464 kubelet[2811]: I0430 00:37:54.765290 2811 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 30 00:37:54.767913 kubelet[2811]: E0430 00:37:54.767859 2811 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.3.3-a-c388cdac8a\" not found" Apr 30 00:37:54.826969 systemd[1]: Created slice kubepods-burstable-pod32bb5359c77dfc7f1bdbf25ff325e9fe.slice - libcontainer container kubepods-burstable-pod32bb5359c77dfc7f1bdbf25ff325e9fe.slice. Apr 30 00:37:54.839265 systemd[1]: Created slice kubepods-burstable-pod683578e250f504b1b7dc6d6c70c05799.slice - libcontainer container kubepods-burstable-pod683578e250f504b1b7dc6d6c70c05799.slice. Apr 30 00:37:54.850816 systemd[1]: Created slice kubepods-burstable-pod5fd1fbb0e1dee884d8f7355ebe1430c6.slice - libcontainer container kubepods-burstable-pod5fd1fbb0e1dee884d8f7355ebe1430c6.slice. Apr 30 00:37:54.866268 kubelet[2811]: I0430 00:37:54.866192 2811 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.866555 kubelet[2811]: E0430 00:37:54.866529 2811 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.20.12:6443/api/v1/nodes\": dial tcp 10.200.20.12:6443: connect: connection refused" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.876116 kubelet[2811]: E0430 00:37:54.875203 2811 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.3-a-c388cdac8a?timeout=10s\": dial tcp 10.200.20.12:6443: connect: connection refused" interval="400ms" Apr 30 00:37:54.972669 kubelet[2811]: I0430 00:37:54.972515 2811 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5fd1fbb0e1dee884d8f7355ebe1430c6-kubeconfig\") pod \"kube-scheduler-ci-4081.3.3-a-c388cdac8a\" (UID: \"5fd1fbb0e1dee884d8f7355ebe1430c6\") " pod="kube-system/kube-scheduler-ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.972669 kubelet[2811]: I0430 00:37:54.972558 2811 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/32bb5359c77dfc7f1bdbf25ff325e9fe-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.3-a-c388cdac8a\" (UID: \"32bb5359c77dfc7f1bdbf25ff325e9fe\") " pod="kube-system/kube-apiserver-ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.972669 kubelet[2811]: I0430 00:37:54.972583 2811 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.972669 kubelet[2811]: I0430 00:37:54.972598 2811 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.972669 kubelet[2811]: I0430 00:37:54.972614 2811 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.972900 kubelet[2811]: I0430 00:37:54.972630 2811 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/32bb5359c77dfc7f1bdbf25ff325e9fe-ca-certs\") pod \"kube-apiserver-ci-4081.3.3-a-c388cdac8a\" (UID: \"32bb5359c77dfc7f1bdbf25ff325e9fe\") " pod="kube-system/kube-apiserver-ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.972900 kubelet[2811]: I0430 00:37:54.972656 2811 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/32bb5359c77dfc7f1bdbf25ff325e9fe-k8s-certs\") pod \"kube-apiserver-ci-4081.3.3-a-c388cdac8a\" (UID: \"32bb5359c77dfc7f1bdbf25ff325e9fe\") " pod="kube-system/kube-apiserver-ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.972900 kubelet[2811]: I0430 00:37:54.972670 2811 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-ca-certs\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:54.972900 kubelet[2811]: I0430 00:37:54.972686 2811 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:55.069473 kubelet[2811]: I0430 00:37:55.069180 2811 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:55.069628 kubelet[2811]: E0430 00:37:55.069526 2811 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.20.12:6443/api/v1/nodes\": dial tcp 10.200.20.12:6443: connect: connection refused" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:55.136948 containerd[1704]: time="2025-04-30T00:37:55.136416616Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.3-a-c388cdac8a,Uid:32bb5359c77dfc7f1bdbf25ff325e9fe,Namespace:kube-system,Attempt:0,}" Apr 30 00:37:55.149446 containerd[1704]: time="2025-04-30T00:37:55.149385626Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.3-a-c388cdac8a,Uid:683578e250f504b1b7dc6d6c70c05799,Namespace:kube-system,Attempt:0,}" Apr 30 00:37:55.154097 containerd[1704]: time="2025-04-30T00:37:55.154035749Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.3-a-c388cdac8a,Uid:5fd1fbb0e1dee884d8f7355ebe1430c6,Namespace:kube-system,Attempt:0,}" Apr 30 00:37:55.276460 kubelet[2811]: E0430 00:37:55.276405 2811 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.3-a-c388cdac8a?timeout=10s\": dial tcp 10.200.20.12:6443: connect: connection refused" interval="800ms" Apr 30 00:37:55.472008 kubelet[2811]: I0430 00:37:55.471797 2811 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:55.472436 kubelet[2811]: E0430 00:37:55.472386 2811 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.20.12:6443/api/v1/nodes\": dial tcp 10.200.20.12:6443: connect: connection refused" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:55.642674 kubelet[2811]: W0430 00:37:55.642584 2811 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.20.12:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.12:6443: connect: connection refused Apr 30 00:37:55.642674 kubelet[2811]: E0430 00:37:55.642636 2811 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.200.20.12:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:55.646407 kubelet[2811]: W0430 00:37:55.646330 2811 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.20.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.12:6443: connect: connection refused Apr 30 00:37:55.646407 kubelet[2811]: E0430 00:37:55.646373 2811 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.200.20.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:55.716209 kubelet[2811]: W0430 00:37:55.716104 2811 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.20.12:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.200.20.12:6443: connect: connection refused Apr 30 00:37:55.716209 kubelet[2811]: E0430 00:37:55.716172 2811 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.200.20.12:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:55.890921 kubelet[2811]: W0430 00:37:55.890805 2811 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.20.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.3-a-c388cdac8a&limit=500&resourceVersion=0": dial tcp 10.200.20.12:6443: connect: connection refused Apr 30 00:37:55.890921 kubelet[2811]: E0430 00:37:55.890878 2811 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.200.20.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.3-a-c388cdac8a&limit=500&resourceVersion=0\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:56.077334 kubelet[2811]: E0430 00:37:56.077275 2811 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.3-a-c388cdac8a?timeout=10s\": dial tcp 10.200.20.12:6443: connect: connection refused" interval="1.6s" Apr 30 00:37:56.274411 kubelet[2811]: I0430 00:37:56.274299 2811 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:56.274919 kubelet[2811]: E0430 00:37:56.274864 2811 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.20.12:6443/api/v1/nodes\": dial tcp 10.200.20.12:6443: connect: connection refused" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:37:56.465452 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2881167665.mount: Deactivated successfully. Apr 30 00:37:56.513443 containerd[1704]: time="2025-04-30T00:37:56.513381903Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 00:37:56.517286 containerd[1704]: time="2025-04-30T00:37:56.516458466Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 00:37:56.519734 containerd[1704]: time="2025-04-30T00:37:56.519609509Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Apr 30 00:37:56.526920 containerd[1704]: time="2025-04-30T00:37:56.526090635Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 00:37:56.530782 containerd[1704]: time="2025-04-30T00:37:56.530725639Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 00:37:56.535680 containerd[1704]: time="2025-04-30T00:37:56.535632284Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 30 00:37:56.538470 containerd[1704]: time="2025-04-30T00:37:56.538104486Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 30 00:37:56.542316 containerd[1704]: time="2025-04-30T00:37:56.542252930Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 00:37:56.543641 containerd[1704]: time="2025-04-30T00:37:56.543127931Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 1.406220315s" Apr 30 00:37:56.549556 containerd[1704]: time="2025-04-30T00:37:56.549502497Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 1.395361468s" Apr 30 00:37:56.557090 containerd[1704]: time="2025-04-30T00:37:56.557031184Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 1.407564918s" Apr 30 00:37:56.812851 kubelet[2811]: E0430 00:37:56.812729 2811 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.200.20.12:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.20.12:6443: connect: connection refused" logger="UnhandledError" Apr 30 00:37:57.155239 containerd[1704]: time="2025-04-30T00:37:57.154514517Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:37:57.157997 containerd[1704]: time="2025-04-30T00:37:57.155152398Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:37:57.157997 containerd[1704]: time="2025-04-30T00:37:57.155278838Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:37:57.157997 containerd[1704]: time="2025-04-30T00:37:57.157755640Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:37:57.160574 containerd[1704]: time="2025-04-30T00:37:57.160327563Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:37:57.160574 containerd[1704]: time="2025-04-30T00:37:57.160403523Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:37:57.160574 containerd[1704]: time="2025-04-30T00:37:57.160430923Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:37:57.160574 containerd[1704]: time="2025-04-30T00:37:57.160522963Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:37:57.160874 containerd[1704]: time="2025-04-30T00:37:57.160734763Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:37:57.160874 containerd[1704]: time="2025-04-30T00:37:57.160843163Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:37:57.160874 containerd[1704]: time="2025-04-30T00:37:57.160907563Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:37:57.161287 containerd[1704]: time="2025-04-30T00:37:57.161192243Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:37:57.185295 systemd[1]: Started cri-containerd-f22c897eb0bd0d77a6cf7af825a839127511b5ebb62284ce1775e4941393c036.scope - libcontainer container f22c897eb0bd0d77a6cf7af825a839127511b5ebb62284ce1775e4941393c036. Apr 30 00:37:57.197288 systemd[1]: Started cri-containerd-1b56e0fbd8f22f75c76d37fd868d3835ddeb725c92ef866ca75382a94995de77.scope - libcontainer container 1b56e0fbd8f22f75c76d37fd868d3835ddeb725c92ef866ca75382a94995de77. Apr 30 00:37:57.199284 systemd[1]: Started cri-containerd-94cafb05ba64dd32932b5afbea36ce78165822594e89c9433f4478318626f6f5.scope - libcontainer container 94cafb05ba64dd32932b5afbea36ce78165822594e89c9433f4478318626f6f5. Apr 30 00:37:57.245910 containerd[1704]: time="2025-04-30T00:37:57.245456244Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.3-a-c388cdac8a,Uid:683578e250f504b1b7dc6d6c70c05799,Namespace:kube-system,Attempt:0,} returns sandbox id \"f22c897eb0bd0d77a6cf7af825a839127511b5ebb62284ce1775e4941393c036\"" Apr 30 00:37:57.252548 containerd[1704]: time="2025-04-30T00:37:57.252391131Z" level=info msg="CreateContainer within sandbox \"f22c897eb0bd0d77a6cf7af825a839127511b5ebb62284ce1775e4941393c036\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 30 00:37:57.253465 containerd[1704]: time="2025-04-30T00:37:57.253349972Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.3-a-c388cdac8a,Uid:32bb5359c77dfc7f1bdbf25ff325e9fe,Namespace:kube-system,Attempt:0,} returns sandbox id \"1b56e0fbd8f22f75c76d37fd868d3835ddeb725c92ef866ca75382a94995de77\"" Apr 30 00:37:57.257934 containerd[1704]: time="2025-04-30T00:37:57.257884136Z" level=info msg="CreateContainer within sandbox \"1b56e0fbd8f22f75c76d37fd868d3835ddeb725c92ef866ca75382a94995de77\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 30 00:37:57.264146 containerd[1704]: time="2025-04-30T00:37:57.264097302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.3-a-c388cdac8a,Uid:5fd1fbb0e1dee884d8f7355ebe1430c6,Namespace:kube-system,Attempt:0,} returns sandbox id \"94cafb05ba64dd32932b5afbea36ce78165822594e89c9433f4478318626f6f5\"" Apr 30 00:37:57.267960 containerd[1704]: time="2025-04-30T00:37:57.267916346Z" level=info msg="CreateContainer within sandbox \"94cafb05ba64dd32932b5afbea36ce78165822594e89c9433f4478318626f6f5\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 30 00:37:57.322981 kubelet[2811]: E0430 00:37:57.322843 2811 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.20.12:6443/api/v1/namespaces/default/events\": dial tcp 10.200.20.12:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.3-a-c388cdac8a.183af19fe89c464a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.3-a-c388cdac8a,UID:ci-4081.3.3-a-c388cdac8a,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.3-a-c388cdac8a,},FirstTimestamp:2025-04-30 00:37:54.664138314 +0000 UTC m=+1.542761360,LastTimestamp:2025-04-30 00:37:54.664138314 +0000 UTC m=+1.542761360,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.3-a-c388cdac8a,}" Apr 30 00:37:57.381049 containerd[1704]: time="2025-04-30T00:37:57.380991774Z" level=info msg="CreateContainer within sandbox \"f22c897eb0bd0d77a6cf7af825a839127511b5ebb62284ce1775e4941393c036\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"38e265257e124abb8c21f9ebec09d2f58979471024f97e471624be98580d0d30\"" Apr 30 00:37:57.381772 containerd[1704]: time="2025-04-30T00:37:57.381736695Z" level=info msg="StartContainer for \"38e265257e124abb8c21f9ebec09d2f58979471024f97e471624be98580d0d30\"" Apr 30 00:37:57.395993 containerd[1704]: time="2025-04-30T00:37:57.395925788Z" level=info msg="CreateContainer within sandbox \"94cafb05ba64dd32932b5afbea36ce78165822594e89c9433f4478318626f6f5\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"508e0ef2699c8d7754986f5f918e40da0e907e6e5d1d49d0be7e1a19fb1a33e4\"" Apr 30 00:37:57.396583 containerd[1704]: time="2025-04-30T00:37:57.396536669Z" level=info msg="StartContainer for \"508e0ef2699c8d7754986f5f918e40da0e907e6e5d1d49d0be7e1a19fb1a33e4\"" Apr 30 00:37:57.400683 containerd[1704]: time="2025-04-30T00:37:57.400629913Z" level=info msg="CreateContainer within sandbox \"1b56e0fbd8f22f75c76d37fd868d3835ddeb725c92ef866ca75382a94995de77\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"3d3d48f7e53f4476d9c9a17aeae6378ece58826df48f053b6126b2387c895e2c\"" Apr 30 00:37:57.401299 containerd[1704]: time="2025-04-30T00:37:57.401267793Z" level=info msg="StartContainer for \"3d3d48f7e53f4476d9c9a17aeae6378ece58826df48f053b6126b2387c895e2c\"" Apr 30 00:37:57.410389 systemd[1]: Started cri-containerd-38e265257e124abb8c21f9ebec09d2f58979471024f97e471624be98580d0d30.scope - libcontainer container 38e265257e124abb8c21f9ebec09d2f58979471024f97e471624be98580d0d30. Apr 30 00:37:57.443414 systemd[1]: Started cri-containerd-508e0ef2699c8d7754986f5f918e40da0e907e6e5d1d49d0be7e1a19fb1a33e4.scope - libcontainer container 508e0ef2699c8d7754986f5f918e40da0e907e6e5d1d49d0be7e1a19fb1a33e4. Apr 30 00:37:57.452559 systemd[1]: Started cri-containerd-3d3d48f7e53f4476d9c9a17aeae6378ece58826df48f053b6126b2387c895e2c.scope - libcontainer container 3d3d48f7e53f4476d9c9a17aeae6378ece58826df48f053b6126b2387c895e2c. Apr 30 00:37:57.472894 containerd[1704]: time="2025-04-30T00:37:57.472847702Z" level=info msg="StartContainer for \"38e265257e124abb8c21f9ebec09d2f58979471024f97e471624be98580d0d30\" returns successfully" Apr 30 00:37:57.525400 containerd[1704]: time="2025-04-30T00:37:57.525341192Z" level=info msg="StartContainer for \"508e0ef2699c8d7754986f5f918e40da0e907e6e5d1d49d0be7e1a19fb1a33e4\" returns successfully" Apr 30 00:37:57.538954 containerd[1704]: time="2025-04-30T00:37:57.538899125Z" level=info msg="StartContainer for \"3d3d48f7e53f4476d9c9a17aeae6378ece58826df48f053b6126b2387c895e2c\" returns successfully" Apr 30 00:37:57.879183 kubelet[2811]: I0430 00:37:57.876961 2811 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:00.335171 kubelet[2811]: E0430 00:38:00.335012 2811 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081.3.3-a-c388cdac8a\" not found" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:00.484527 kubelet[2811]: I0430 00:38:00.483104 2811 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:00.662156 kubelet[2811]: I0430 00:38:00.662115 2811 apiserver.go:52] "Watching apiserver" Apr 30 00:38:00.671836 kubelet[2811]: I0430 00:38:00.671787 2811 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 30 00:38:02.536003 kubelet[2811]: W0430 00:38:02.535952 2811 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 30 00:38:04.571761 systemd[1]: Reloading requested from client PID 3090 ('systemctl') (unit session-9.scope)... Apr 30 00:38:04.571780 systemd[1]: Reloading... Apr 30 00:38:04.688120 zram_generator::config[3133]: No configuration found. Apr 30 00:38:04.839545 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 00:38:04.933947 systemd[1]: Reloading finished in 361 ms. Apr 30 00:38:04.977412 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:38:04.993568 systemd[1]: kubelet.service: Deactivated successfully. Apr 30 00:38:04.993829 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:38:04.993887 systemd[1]: kubelet.service: Consumed 1.987s CPU time, 116.4M memory peak, 0B memory swap peak. Apr 30 00:38:04.998373 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 00:38:05.151270 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 00:38:05.161223 (kubelet)[3195]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 30 00:38:05.213246 kubelet[3195]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 00:38:05.213246 kubelet[3195]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Apr 30 00:38:05.213246 kubelet[3195]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 00:38:05.213246 kubelet[3195]: I0430 00:38:05.212778 3195 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 30 00:38:05.220460 kubelet[3195]: I0430 00:38:05.220419 3195 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Apr 30 00:38:05.221865 kubelet[3195]: I0430 00:38:05.220793 3195 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 30 00:38:05.221865 kubelet[3195]: I0430 00:38:05.221153 3195 server.go:929] "Client rotation is on, will bootstrap in background" Apr 30 00:38:05.225128 kubelet[3195]: I0430 00:38:05.223496 3195 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Apr 30 00:38:05.226133 kubelet[3195]: I0430 00:38:05.225926 3195 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 30 00:38:05.230537 kubelet[3195]: E0430 00:38:05.230422 3195 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 30 00:38:05.230537 kubelet[3195]: I0430 00:38:05.230458 3195 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 30 00:38:05.234499 kubelet[3195]: I0430 00:38:05.234455 3195 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 30 00:38:05.234658 kubelet[3195]: I0430 00:38:05.234580 3195 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Apr 30 00:38:05.234696 kubelet[3195]: I0430 00:38:05.234662 3195 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 30 00:38:05.234879 kubelet[3195]: I0430 00:38:05.234688 3195 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.3-a-c388cdac8a","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 30 00:38:05.235707 kubelet[3195]: I0430 00:38:05.234885 3195 topology_manager.go:138] "Creating topology manager with none policy" Apr 30 00:38:05.235707 kubelet[3195]: I0430 00:38:05.234901 3195 container_manager_linux.go:300] "Creating device plugin manager" Apr 30 00:38:05.235707 kubelet[3195]: I0430 00:38:05.234934 3195 state_mem.go:36] "Initialized new in-memory state store" Apr 30 00:38:05.235707 kubelet[3195]: I0430 00:38:05.235054 3195 kubelet.go:408] "Attempting to sync node with API server" Apr 30 00:38:05.235707 kubelet[3195]: I0430 00:38:05.235088 3195 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 30 00:38:05.235707 kubelet[3195]: I0430 00:38:05.235116 3195 kubelet.go:314] "Adding apiserver pod source" Apr 30 00:38:05.235707 kubelet[3195]: I0430 00:38:05.235126 3195 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 30 00:38:05.239455 kubelet[3195]: I0430 00:38:05.239411 3195 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 30 00:38:05.241273 kubelet[3195]: I0430 00:38:05.240180 3195 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Apr 30 00:38:05.241273 kubelet[3195]: I0430 00:38:05.240907 3195 server.go:1269] "Started kubelet" Apr 30 00:38:05.247776 kubelet[3195]: I0430 00:38:05.247721 3195 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Apr 30 00:38:05.249097 kubelet[3195]: I0430 00:38:05.249035 3195 server.go:460] "Adding debug handlers to kubelet server" Apr 30 00:38:05.250908 kubelet[3195]: I0430 00:38:05.250831 3195 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 30 00:38:05.251399 kubelet[3195]: I0430 00:38:05.251365 3195 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 30 00:38:05.254486 kubelet[3195]: I0430 00:38:05.254455 3195 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 30 00:38:05.262329 kubelet[3195]: I0430 00:38:05.262278 3195 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 30 00:38:05.262573 kubelet[3195]: I0430 00:38:05.262553 3195 volume_manager.go:289] "Starting Kubelet Volume Manager" Apr 30 00:38:05.262996 kubelet[3195]: E0430 00:38:05.262955 3195 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.3-a-c388cdac8a\" not found" Apr 30 00:38:05.264796 kubelet[3195]: I0430 00:38:05.264768 3195 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 30 00:38:05.265138 kubelet[3195]: I0430 00:38:05.265124 3195 reconciler.go:26] "Reconciler: start to sync state" Apr 30 00:38:05.268431 kubelet[3195]: I0430 00:38:05.268249 3195 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Apr 30 00:38:05.268571 kubelet[3195]: I0430 00:38:05.268494 3195 factory.go:221] Registration of the systemd container factory successfully Apr 30 00:38:05.268781 kubelet[3195]: I0430 00:38:05.268615 3195 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 30 00:38:05.273004 kubelet[3195]: I0430 00:38:05.272871 3195 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Apr 30 00:38:05.274206 kubelet[3195]: I0430 00:38:05.273429 3195 status_manager.go:217] "Starting to sync pod status with apiserver" Apr 30 00:38:05.274206 kubelet[3195]: I0430 00:38:05.273465 3195 kubelet.go:2321] "Starting kubelet main sync loop" Apr 30 00:38:05.274206 kubelet[3195]: E0430 00:38:05.273521 3195 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 30 00:38:05.284294 kubelet[3195]: E0430 00:38:05.284245 3195 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 30 00:38:05.298116 kubelet[3195]: I0430 00:38:05.296178 3195 factory.go:221] Registration of the containerd container factory successfully Apr 30 00:38:05.364272 kubelet[3195]: I0430 00:38:05.364240 3195 cpu_manager.go:214] "Starting CPU manager" policy="none" Apr 30 00:38:05.364272 kubelet[3195]: I0430 00:38:05.364260 3195 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Apr 30 00:38:05.364272 kubelet[3195]: I0430 00:38:05.364284 3195 state_mem.go:36] "Initialized new in-memory state store" Apr 30 00:38:05.364461 kubelet[3195]: I0430 00:38:05.364434 3195 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 30 00:38:05.364461 kubelet[3195]: I0430 00:38:05.364445 3195 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 30 00:38:05.364514 kubelet[3195]: I0430 00:38:05.364464 3195 policy_none.go:49] "None policy: Start" Apr 30 00:38:05.365637 kubelet[3195]: I0430 00:38:05.365428 3195 memory_manager.go:170] "Starting memorymanager" policy="None" Apr 30 00:38:05.365637 kubelet[3195]: I0430 00:38:05.365487 3195 state_mem.go:35] "Initializing new in-memory state store" Apr 30 00:38:05.365955 kubelet[3195]: I0430 00:38:05.365921 3195 state_mem.go:75] "Updated machine memory state" Apr 30 00:38:05.370758 kubelet[3195]: I0430 00:38:05.370728 3195 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 30 00:38:05.372141 kubelet[3195]: I0430 00:38:05.372107 3195 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 30 00:38:05.372242 kubelet[3195]: I0430 00:38:05.372134 3195 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 30 00:38:05.376381 kubelet[3195]: I0430 00:38:05.376313 3195 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 30 00:38:05.389275 kubelet[3195]: W0430 00:38:05.388338 3195 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 30 00:38:05.396863 kubelet[3195]: W0430 00:38:05.396821 3195 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 30 00:38:05.398062 kubelet[3195]: W0430 00:38:05.398028 3195 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 30 00:38:05.398196 kubelet[3195]: E0430 00:38:05.398111 3195 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4081.3.3-a-c388cdac8a\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.465805 kubelet[3195]: I0430 00:38:05.465454 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-ca-certs\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.465805 kubelet[3195]: I0430 00:38:05.465496 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.465805 kubelet[3195]: I0430 00:38:05.465521 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5fd1fbb0e1dee884d8f7355ebe1430c6-kubeconfig\") pod \"kube-scheduler-ci-4081.3.3-a-c388cdac8a\" (UID: \"5fd1fbb0e1dee884d8f7355ebe1430c6\") " pod="kube-system/kube-scheduler-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.465805 kubelet[3195]: I0430 00:38:05.465557 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/32bb5359c77dfc7f1bdbf25ff325e9fe-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.3-a-c388cdac8a\" (UID: \"32bb5359c77dfc7f1bdbf25ff325e9fe\") " pod="kube-system/kube-apiserver-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.465805 kubelet[3195]: I0430 00:38:05.465576 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.466003 kubelet[3195]: I0430 00:38:05.465597 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.466003 kubelet[3195]: I0430 00:38:05.465613 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/683578e250f504b1b7dc6d6c70c05799-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.3-a-c388cdac8a\" (UID: \"683578e250f504b1b7dc6d6c70c05799\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.466003 kubelet[3195]: I0430 00:38:05.465628 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/32bb5359c77dfc7f1bdbf25ff325e9fe-ca-certs\") pod \"kube-apiserver-ci-4081.3.3-a-c388cdac8a\" (UID: \"32bb5359c77dfc7f1bdbf25ff325e9fe\") " pod="kube-system/kube-apiserver-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.466003 kubelet[3195]: I0430 00:38:05.465643 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/32bb5359c77dfc7f1bdbf25ff325e9fe-k8s-certs\") pod \"kube-apiserver-ci-4081.3.3-a-c388cdac8a\" (UID: \"32bb5359c77dfc7f1bdbf25ff325e9fe\") " pod="kube-system/kube-apiserver-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.487684 kubelet[3195]: I0430 00:38:05.487637 3195 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.511855 kubelet[3195]: I0430 00:38:05.511020 3195 kubelet_node_status.go:111] "Node was previously registered" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:05.511855 kubelet[3195]: I0430 00:38:05.511149 3195 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:06.236474 kubelet[3195]: I0430 00:38:06.236129 3195 apiserver.go:52] "Watching apiserver" Apr 30 00:38:06.267258 kubelet[3195]: I0430 00:38:06.266043 3195 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 30 00:38:06.394145 kubelet[3195]: W0430 00:38:06.394104 3195 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 30 00:38:06.394315 kubelet[3195]: E0430 00:38:06.394178 3195 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4081.3.3-a-c388cdac8a\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:06.474425 kubelet[3195]: I0430 00:38:06.474135 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.3.3-a-c388cdac8a" podStartSLOduration=1.474105186 podStartE2EDuration="1.474105186s" podCreationTimestamp="2025-04-30 00:38:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 00:38:06.420744616 +0000 UTC m=+1.255970784" watchObservedRunningTime="2025-04-30 00:38:06.474105186 +0000 UTC m=+1.309331354" Apr 30 00:38:06.695783 kubelet[3195]: I0430 00:38:06.695696 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.3.3-a-c388cdac8a" podStartSLOduration=1.695681207 podStartE2EDuration="1.695681207s" podCreationTimestamp="2025-04-30 00:38:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 00:38:06.643457079 +0000 UTC m=+1.478683247" watchObservedRunningTime="2025-04-30 00:38:06.695681207 +0000 UTC m=+1.530907375" Apr 30 00:38:09.061886 kubelet[3195]: I0430 00:38:09.061844 3195 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 30 00:38:09.063561 containerd[1704]: time="2025-04-30T00:38:09.062695328Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 30 00:38:09.065553 kubelet[3195]: I0430 00:38:09.063095 3195 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 30 00:38:09.695211 kubelet[3195]: I0430 00:38:09.695162 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgr54\" (UniqueName: \"kubernetes.io/projected/a4c4a2a4-0696-4fca-92ef-d4ded65b3d36-kube-api-access-bgr54\") pod \"kube-proxy-8mgkk\" (UID: \"a4c4a2a4-0696-4fca-92ef-d4ded65b3d36\") " pod="kube-system/kube-proxy-8mgkk" Apr 30 00:38:09.696246 kubelet[3195]: I0430 00:38:09.695461 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a4c4a2a4-0696-4fca-92ef-d4ded65b3d36-xtables-lock\") pod \"kube-proxy-8mgkk\" (UID: \"a4c4a2a4-0696-4fca-92ef-d4ded65b3d36\") " pod="kube-system/kube-proxy-8mgkk" Apr 30 00:38:09.696246 kubelet[3195]: I0430 00:38:09.695489 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a4c4a2a4-0696-4fca-92ef-d4ded65b3d36-lib-modules\") pod \"kube-proxy-8mgkk\" (UID: \"a4c4a2a4-0696-4fca-92ef-d4ded65b3d36\") " pod="kube-system/kube-proxy-8mgkk" Apr 30 00:38:09.697277 systemd[1]: Created slice kubepods-besteffort-poda4c4a2a4_0696_4fca_92ef_d4ded65b3d36.slice - libcontainer container kubepods-besteffort-poda4c4a2a4_0696_4fca_92ef_d4ded65b3d36.slice. Apr 30 00:38:09.702254 kubelet[3195]: I0430 00:38:09.701681 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/a4c4a2a4-0696-4fca-92ef-d4ded65b3d36-kube-proxy\") pod \"kube-proxy-8mgkk\" (UID: \"a4c4a2a4-0696-4fca-92ef-d4ded65b3d36\") " pod="kube-system/kube-proxy-8mgkk" Apr 30 00:38:10.026135 containerd[1704]: time="2025-04-30T00:38:10.025644707Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8mgkk,Uid:a4c4a2a4-0696-4fca-92ef-d4ded65b3d36,Namespace:kube-system,Attempt:0,}" Apr 30 00:38:10.111888 containerd[1704]: time="2025-04-30T00:38:10.111770268Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:38:10.111888 containerd[1704]: time="2025-04-30T00:38:10.111913748Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:38:10.112625 containerd[1704]: time="2025-04-30T00:38:10.111948548Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:10.112625 containerd[1704]: time="2025-04-30T00:38:10.112113468Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:10.137324 systemd[1]: Started cri-containerd-5419111af8f5455585f4441f21e126fcb69773cad653d4cd01cd9eeaf8ceecd2.scope - libcontainer container 5419111af8f5455585f4441f21e126fcb69773cad653d4cd01cd9eeaf8ceecd2. Apr 30 00:38:10.174334 containerd[1704]: time="2025-04-30T00:38:10.174266366Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8mgkk,Uid:a4c4a2a4-0696-4fca-92ef-d4ded65b3d36,Namespace:kube-system,Attempt:0,} returns sandbox id \"5419111af8f5455585f4441f21e126fcb69773cad653d4cd01cd9eeaf8ceecd2\"" Apr 30 00:38:10.183980 containerd[1704]: time="2025-04-30T00:38:10.183577055Z" level=info msg="CreateContainer within sandbox \"5419111af8f5455585f4441f21e126fcb69773cad653d4cd01cd9eeaf8ceecd2\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 30 00:38:10.239299 systemd[1]: Created slice kubepods-besteffort-pod9a073e74_e506_4d94_a42d_a4b3fc880e0e.slice - libcontainer container kubepods-besteffort-pod9a073e74_e506_4d94_a42d_a4b3fc880e0e.slice. Apr 30 00:38:10.251794 containerd[1704]: time="2025-04-30T00:38:10.251722358Z" level=info msg="CreateContainer within sandbox \"5419111af8f5455585f4441f21e126fcb69773cad653d4cd01cd9eeaf8ceecd2\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"9c70114849bfbafca6fc8d80aed9cca96d49ca006b059274f0b05e39885c3f11\"" Apr 30 00:38:10.253707 containerd[1704]: time="2025-04-30T00:38:10.253482280Z" level=info msg="StartContainer for \"9c70114849bfbafca6fc8d80aed9cca96d49ca006b059274f0b05e39885c3f11\"" Apr 30 00:38:10.295343 systemd[1]: Started cri-containerd-9c70114849bfbafca6fc8d80aed9cca96d49ca006b059274f0b05e39885c3f11.scope - libcontainer container 9c70114849bfbafca6fc8d80aed9cca96d49ca006b059274f0b05e39885c3f11. Apr 30 00:38:10.305916 kubelet[3195]: I0430 00:38:10.305808 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvw87\" (UniqueName: \"kubernetes.io/projected/9a073e74-e506-4d94-a42d-a4b3fc880e0e-kube-api-access-xvw87\") pod \"tigera-operator-6f6897fdc5-r6lzw\" (UID: \"9a073e74-e506-4d94-a42d-a4b3fc880e0e\") " pod="tigera-operator/tigera-operator-6f6897fdc5-r6lzw" Apr 30 00:38:10.305916 kubelet[3195]: I0430 00:38:10.305873 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/9a073e74-e506-4d94-a42d-a4b3fc880e0e-var-lib-calico\") pod \"tigera-operator-6f6897fdc5-r6lzw\" (UID: \"9a073e74-e506-4d94-a42d-a4b3fc880e0e\") " pod="tigera-operator/tigera-operator-6f6897fdc5-r6lzw" Apr 30 00:38:10.336373 containerd[1704]: time="2025-04-30T00:38:10.336214757Z" level=info msg="StartContainer for \"9c70114849bfbafca6fc8d80aed9cca96d49ca006b059274f0b05e39885c3f11\" returns successfully" Apr 30 00:38:10.395948 kubelet[3195]: I0430 00:38:10.395728 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-8mgkk" podStartSLOduration=1.395705853 podStartE2EDuration="1.395705853s" podCreationTimestamp="2025-04-30 00:38:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 00:38:10.395608573 +0000 UTC m=+5.230834741" watchObservedRunningTime="2025-04-30 00:38:10.395705853 +0000 UTC m=+5.230932021" Apr 30 00:38:10.552237 containerd[1704]: time="2025-04-30T00:38:10.551569398Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6f6897fdc5-r6lzw,Uid:9a073e74-e506-4d94-a42d-a4b3fc880e0e,Namespace:tigera-operator,Attempt:0,}" Apr 30 00:38:10.617164 containerd[1704]: time="2025-04-30T00:38:10.616936979Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:38:10.617164 containerd[1704]: time="2025-04-30T00:38:10.617015379Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:38:10.617164 containerd[1704]: time="2025-04-30T00:38:10.617026859Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:10.617481 containerd[1704]: time="2025-04-30T00:38:10.617155980Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:10.636852 systemd[1]: Started cri-containerd-26a1926bcf3081849fea868cd8b769c1270d554639d0a25189620a593dcd9949.scope - libcontainer container 26a1926bcf3081849fea868cd8b769c1270d554639d0a25189620a593dcd9949. Apr 30 00:38:10.673933 containerd[1704]: time="2025-04-30T00:38:10.673865193Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6f6897fdc5-r6lzw,Uid:9a073e74-e506-4d94-a42d-a4b3fc880e0e,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"26a1926bcf3081849fea868cd8b769c1270d554639d0a25189620a593dcd9949\"" Apr 30 00:38:10.676403 containerd[1704]: time="2025-04-30T00:38:10.676357035Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\"" Apr 30 00:38:12.598432 sudo[2215]: pam_unix(sudo:session): session closed for user root Apr 30 00:38:12.613966 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1790566848.mount: Deactivated successfully. Apr 30 00:38:12.664333 sshd[2212]: pam_unix(sshd:session): session closed for user core Apr 30 00:38:12.668627 systemd[1]: sshd@6-10.200.20.12:22-10.200.16.10:34598.service: Deactivated successfully. Apr 30 00:38:12.671609 systemd[1]: session-9.scope: Deactivated successfully. Apr 30 00:38:12.671845 systemd[1]: session-9.scope: Consumed 7.871s CPU time, 153.6M memory peak, 0B memory swap peak. Apr 30 00:38:12.673225 systemd-logind[1668]: Session 9 logged out. Waiting for processes to exit. Apr 30 00:38:12.674756 systemd-logind[1668]: Removed session 9. Apr 30 00:38:13.449665 containerd[1704]: time="2025-04-30T00:38:13.449604448Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:13.454837 containerd[1704]: time="2025-04-30T00:38:13.454574373Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.7: active requests=0, bytes read=19323084" Apr 30 00:38:13.469378 containerd[1704]: time="2025-04-30T00:38:13.468582188Z" level=info msg="ImageCreate event name:\"sha256:27f7c2cfac802523e44ecd16453a4cc992f6c7d610c13054f2715a7cb4370565\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:13.478754 containerd[1704]: time="2025-04-30T00:38:13.478686799Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:13.479862 containerd[1704]: time="2025-04-30T00:38:13.479289560Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.7\" with image id \"sha256:27f7c2cfac802523e44ecd16453a4cc992f6c7d610c13054f2715a7cb4370565\", repo tag \"quay.io/tigera/operator:v1.36.7\", repo digest \"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\", size \"19319079\" in 2.802895085s" Apr 30 00:38:13.479862 containerd[1704]: time="2025-04-30T00:38:13.479331640Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\" returns image reference \"sha256:27f7c2cfac802523e44ecd16453a4cc992f6c7d610c13054f2715a7cb4370565\"" Apr 30 00:38:13.482944 containerd[1704]: time="2025-04-30T00:38:13.482888084Z" level=info msg="CreateContainer within sandbox \"26a1926bcf3081849fea868cd8b769c1270d554639d0a25189620a593dcd9949\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Apr 30 00:38:13.539654 containerd[1704]: time="2025-04-30T00:38:13.539598906Z" level=info msg="CreateContainer within sandbox \"26a1926bcf3081849fea868cd8b769c1270d554639d0a25189620a593dcd9949\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"84f4fa54569207316cfbe328f0ac6b820f597eb2a2a1ff45678a7f22d1e57501\"" Apr 30 00:38:13.541256 containerd[1704]: time="2025-04-30T00:38:13.541180187Z" level=info msg="StartContainer for \"84f4fa54569207316cfbe328f0ac6b820f597eb2a2a1ff45678a7f22d1e57501\"" Apr 30 00:38:13.572278 systemd[1]: Started cri-containerd-84f4fa54569207316cfbe328f0ac6b820f597eb2a2a1ff45678a7f22d1e57501.scope - libcontainer container 84f4fa54569207316cfbe328f0ac6b820f597eb2a2a1ff45678a7f22d1e57501. Apr 30 00:38:13.604475 containerd[1704]: time="2025-04-30T00:38:13.604399016Z" level=info msg="StartContainer for \"84f4fa54569207316cfbe328f0ac6b820f597eb2a2a1ff45678a7f22d1e57501\" returns successfully" Apr 30 00:38:15.290041 kubelet[3195]: I0430 00:38:15.289881 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6f6897fdc5-r6lzw" podStartSLOduration=3.484826042 podStartE2EDuration="6.289859769s" podCreationTimestamp="2025-04-30 00:38:09 +0000 UTC" firstStartedPulling="2025-04-30 00:38:10.675417274 +0000 UTC m=+5.510643442" lastFinishedPulling="2025-04-30 00:38:13.480451041 +0000 UTC m=+8.315677169" observedRunningTime="2025-04-30 00:38:14.558883854 +0000 UTC m=+9.394110022" watchObservedRunningTime="2025-04-30 00:38:15.289859769 +0000 UTC m=+10.125085937" Apr 30 00:38:18.754132 systemd[1]: Created slice kubepods-besteffort-pod67020db6_c4b9_4da2_be40_961ad42451f6.slice - libcontainer container kubepods-besteffort-pod67020db6_c4b9_4da2_be40_961ad42451f6.slice. Apr 30 00:38:18.857619 kubelet[3195]: I0430 00:38:18.857352 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfsjk\" (UniqueName: \"kubernetes.io/projected/67020db6-c4b9-4da2-be40-961ad42451f6-kube-api-access-kfsjk\") pod \"calico-typha-5dd8879757-zqhv7\" (UID: \"67020db6-c4b9-4da2-be40-961ad42451f6\") " pod="calico-system/calico-typha-5dd8879757-zqhv7" Apr 30 00:38:18.857619 kubelet[3195]: I0430 00:38:18.857408 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67020db6-c4b9-4da2-be40-961ad42451f6-tigera-ca-bundle\") pod \"calico-typha-5dd8879757-zqhv7\" (UID: \"67020db6-c4b9-4da2-be40-961ad42451f6\") " pod="calico-system/calico-typha-5dd8879757-zqhv7" Apr 30 00:38:18.857619 kubelet[3195]: I0430 00:38:18.857430 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/67020db6-c4b9-4da2-be40-961ad42451f6-typha-certs\") pod \"calico-typha-5dd8879757-zqhv7\" (UID: \"67020db6-c4b9-4da2-be40-961ad42451f6\") " pod="calico-system/calico-typha-5dd8879757-zqhv7" Apr 30 00:38:19.046478 systemd[1]: Created slice kubepods-besteffort-pod7570c5da_24ed_4e7c_9b5f_ed2619cd28d9.slice - libcontainer container kubepods-besteffort-pod7570c5da_24ed_4e7c_9b5f_ed2619cd28d9.slice. Apr 30 00:38:19.063324 containerd[1704]: time="2025-04-30T00:38:19.062904752Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5dd8879757-zqhv7,Uid:67020db6-c4b9-4da2-be40-961ad42451f6,Namespace:calico-system,Attempt:0,}" Apr 30 00:38:19.126832 containerd[1704]: time="2025-04-30T00:38:19.126419582Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:38:19.126832 containerd[1704]: time="2025-04-30T00:38:19.126491662Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:38:19.126832 containerd[1704]: time="2025-04-30T00:38:19.126538822Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:19.126832 containerd[1704]: time="2025-04-30T00:38:19.126623262Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:19.161446 kubelet[3195]: I0430 00:38:19.161322 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-lib-modules\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.162184 kubelet[3195]: I0430 00:38:19.161899 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-cni-log-dir\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.162184 kubelet[3195]: I0430 00:38:19.162047 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-node-certs\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.162184 kubelet[3195]: I0430 00:38:19.162104 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-cni-bin-dir\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.162184 kubelet[3195]: I0430 00:38:19.162129 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-var-lib-calico\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.162184 kubelet[3195]: I0430 00:38:19.162149 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-tigera-ca-bundle\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.163502 kubelet[3195]: I0430 00:38:19.162285 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-flexvol-driver-host\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.163502 kubelet[3195]: I0430 00:38:19.162315 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd25f\" (UniqueName: \"kubernetes.io/projected/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-kube-api-access-qd25f\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.163502 kubelet[3195]: I0430 00:38:19.162450 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-cni-net-dir\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.163502 kubelet[3195]: I0430 00:38:19.162472 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-xtables-lock\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.163502 kubelet[3195]: I0430 00:38:19.162487 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-policysync\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.163633 kubelet[3195]: I0430 00:38:19.162613 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/7570c5da-24ed-4e7c-9b5f-ed2619cd28d9-var-run-calico\") pod \"calico-node-9jgff\" (UID: \"7570c5da-24ed-4e7c-9b5f-ed2619cd28d9\") " pod="calico-system/calico-node-9jgff" Apr 30 00:38:19.166324 systemd[1]: Started cri-containerd-b0628cd2226d9ea1ec6b12cbae1d3c0d5097b82551c22a00cfd43bd30ea6b7b5.scope - libcontainer container b0628cd2226d9ea1ec6b12cbae1d3c0d5097b82551c22a00cfd43bd30ea6b7b5. Apr 30 00:38:19.210044 containerd[1704]: time="2025-04-30T00:38:19.209181432Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5dd8879757-zqhv7,Uid:67020db6-c4b9-4da2-be40-961ad42451f6,Namespace:calico-system,Attempt:0,} returns sandbox id \"b0628cd2226d9ea1ec6b12cbae1d3c0d5097b82551c22a00cfd43bd30ea6b7b5\"" Apr 30 00:38:19.212931 containerd[1704]: time="2025-04-30T00:38:19.212873676Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\"" Apr 30 00:38:19.272550 kubelet[3195]: E0430 00:38:19.272133 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.272550 kubelet[3195]: W0430 00:38:19.272259 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.272550 kubelet[3195]: E0430 00:38:19.272295 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.274560 kubelet[3195]: E0430 00:38:19.274343 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.274560 kubelet[3195]: W0430 00:38:19.274369 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.275161 kubelet[3195]: E0430 00:38:19.275005 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.275161 kubelet[3195]: W0430 00:38:19.275083 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.276844 kubelet[3195]: E0430 00:38:19.276720 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.276844 kubelet[3195]: W0430 00:38:19.276741 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.277112 kubelet[3195]: E0430 00:38:19.276998 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.277162 kubelet[3195]: E0430 00:38:19.277139 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.277286 kubelet[3195]: E0430 00:38:19.277210 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.277953 kubelet[3195]: E0430 00:38:19.277840 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.277953 kubelet[3195]: W0430 00:38:19.277859 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.277953 kubelet[3195]: E0430 00:38:19.277882 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.278103 kubelet[3195]: E0430 00:38:19.278040 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.278103 kubelet[3195]: W0430 00:38:19.278047 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.278161 kubelet[3195]: E0430 00:38:19.278142 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.279218 kubelet[3195]: E0430 00:38:19.279044 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.279218 kubelet[3195]: W0430 00:38:19.279123 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.280943 kubelet[3195]: E0430 00:38:19.280692 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.280943 kubelet[3195]: W0430 00:38:19.280727 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.280943 kubelet[3195]: E0430 00:38:19.280762 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.285864 kubelet[3195]: E0430 00:38:19.282368 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.286477 kubelet[3195]: W0430 00:38:19.286150 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.286477 kubelet[3195]: E0430 00:38:19.286349 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.286477 kubelet[3195]: E0430 00:38:19.286123 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.365310 kubelet[3195]: E0430 00:38:19.365274 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.365571 kubelet[3195]: W0430 00:38:19.365476 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.365571 kubelet[3195]: E0430 00:38:19.365507 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.434593 kubelet[3195]: E0430 00:38:19.434222 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:19.461744 kubelet[3195]: E0430 00:38:19.461668 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.461744 kubelet[3195]: W0430 00:38:19.461692 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.461744 kubelet[3195]: E0430 00:38:19.461712 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.461945 kubelet[3195]: E0430 00:38:19.461908 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.461945 kubelet[3195]: W0430 00:38:19.461917 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.461945 kubelet[3195]: E0430 00:38:19.461926 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.462179 kubelet[3195]: E0430 00:38:19.462160 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.462179 kubelet[3195]: W0430 00:38:19.462176 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.462239 kubelet[3195]: E0430 00:38:19.462186 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.462393 kubelet[3195]: E0430 00:38:19.462375 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.462393 kubelet[3195]: W0430 00:38:19.462390 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.462445 kubelet[3195]: E0430 00:38:19.462399 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.462628 kubelet[3195]: E0430 00:38:19.462606 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.462628 kubelet[3195]: W0430 00:38:19.462621 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.462628 kubelet[3195]: E0430 00:38:19.462629 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.462814 kubelet[3195]: E0430 00:38:19.462795 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.462814 kubelet[3195]: W0430 00:38:19.462810 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.462869 kubelet[3195]: E0430 00:38:19.462819 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.462993 kubelet[3195]: E0430 00:38:19.462975 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.462993 kubelet[3195]: W0430 00:38:19.462989 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.463037 kubelet[3195]: E0430 00:38:19.463009 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.463221 kubelet[3195]: E0430 00:38:19.463200 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.463221 kubelet[3195]: W0430 00:38:19.463214 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.463221 kubelet[3195]: E0430 00:38:19.463222 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.463424 kubelet[3195]: E0430 00:38:19.463402 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.463424 kubelet[3195]: W0430 00:38:19.463420 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.463475 kubelet[3195]: E0430 00:38:19.463430 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.463609 kubelet[3195]: E0430 00:38:19.463591 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.463639 kubelet[3195]: W0430 00:38:19.463617 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.463639 kubelet[3195]: E0430 00:38:19.463628 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.463839 kubelet[3195]: E0430 00:38:19.463819 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.463839 kubelet[3195]: W0430 00:38:19.463833 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.463899 kubelet[3195]: E0430 00:38:19.463842 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.464033 kubelet[3195]: E0430 00:38:19.464013 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.464033 kubelet[3195]: W0430 00:38:19.464028 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.464116 kubelet[3195]: E0430 00:38:19.464037 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.464288 kubelet[3195]: E0430 00:38:19.464268 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.464288 kubelet[3195]: W0430 00:38:19.464285 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.464339 kubelet[3195]: E0430 00:38:19.464295 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.464490 kubelet[3195]: E0430 00:38:19.464471 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.464490 kubelet[3195]: W0430 00:38:19.464487 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.464538 kubelet[3195]: E0430 00:38:19.464496 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.464706 kubelet[3195]: E0430 00:38:19.464684 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.464706 kubelet[3195]: W0430 00:38:19.464698 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.464766 kubelet[3195]: E0430 00:38:19.464708 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.464931 kubelet[3195]: E0430 00:38:19.464909 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.464931 kubelet[3195]: W0430 00:38:19.464926 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.464983 kubelet[3195]: E0430 00:38:19.464936 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.465202 kubelet[3195]: E0430 00:38:19.465181 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.465202 kubelet[3195]: W0430 00:38:19.465196 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.465270 kubelet[3195]: E0430 00:38:19.465207 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.465403 kubelet[3195]: E0430 00:38:19.465381 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.465403 kubelet[3195]: W0430 00:38:19.465398 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.465456 kubelet[3195]: E0430 00:38:19.465406 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.465592 kubelet[3195]: E0430 00:38:19.465574 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.465592 kubelet[3195]: W0430 00:38:19.465588 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.465638 kubelet[3195]: E0430 00:38:19.465596 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.465803 kubelet[3195]: E0430 00:38:19.465782 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.465803 kubelet[3195]: W0430 00:38:19.465796 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.465867 kubelet[3195]: E0430 00:38:19.465805 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.466158 kubelet[3195]: E0430 00:38:19.466138 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.466158 kubelet[3195]: W0430 00:38:19.466154 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.466226 kubelet[3195]: E0430 00:38:19.466164 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.466226 kubelet[3195]: I0430 00:38:19.466192 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/633a0f7e-ae5b-47ef-baa9-358c2fe3b117-registration-dir\") pod \"csi-node-driver-shvzf\" (UID: \"633a0f7e-ae5b-47ef-baa9-358c2fe3b117\") " pod="calico-system/csi-node-driver-shvzf" Apr 30 00:38:19.466438 kubelet[3195]: E0430 00:38:19.466417 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.466438 kubelet[3195]: W0430 00:38:19.466433 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.466496 kubelet[3195]: E0430 00:38:19.466451 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.466496 kubelet[3195]: I0430 00:38:19.466466 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw9hz\" (UniqueName: \"kubernetes.io/projected/633a0f7e-ae5b-47ef-baa9-358c2fe3b117-kube-api-access-mw9hz\") pod \"csi-node-driver-shvzf\" (UID: \"633a0f7e-ae5b-47ef-baa9-358c2fe3b117\") " pod="calico-system/csi-node-driver-shvzf" Apr 30 00:38:19.466691 kubelet[3195]: E0430 00:38:19.466669 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.466691 kubelet[3195]: W0430 00:38:19.466685 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.466755 kubelet[3195]: E0430 00:38:19.466700 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.466755 kubelet[3195]: I0430 00:38:19.466717 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/633a0f7e-ae5b-47ef-baa9-358c2fe3b117-varrun\") pod \"csi-node-driver-shvzf\" (UID: \"633a0f7e-ae5b-47ef-baa9-358c2fe3b117\") " pod="calico-system/csi-node-driver-shvzf" Apr 30 00:38:19.466938 kubelet[3195]: E0430 00:38:19.466918 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.466938 kubelet[3195]: W0430 00:38:19.466934 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.466990 kubelet[3195]: E0430 00:38:19.466951 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.467220 kubelet[3195]: E0430 00:38:19.467198 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.467220 kubelet[3195]: W0430 00:38:19.467213 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.467288 kubelet[3195]: E0430 00:38:19.467229 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.467288 kubelet[3195]: I0430 00:38:19.467246 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/633a0f7e-ae5b-47ef-baa9-358c2fe3b117-kubelet-dir\") pod \"csi-node-driver-shvzf\" (UID: \"633a0f7e-ae5b-47ef-baa9-358c2fe3b117\") " pod="calico-system/csi-node-driver-shvzf" Apr 30 00:38:19.467486 kubelet[3195]: E0430 00:38:19.467462 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.467486 kubelet[3195]: W0430 00:38:19.467481 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.467640 kubelet[3195]: E0430 00:38:19.467567 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.467640 kubelet[3195]: I0430 00:38:19.467592 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/633a0f7e-ae5b-47ef-baa9-358c2fe3b117-socket-dir\") pod \"csi-node-driver-shvzf\" (UID: \"633a0f7e-ae5b-47ef-baa9-358c2fe3b117\") " pod="calico-system/csi-node-driver-shvzf" Apr 30 00:38:19.467756 kubelet[3195]: E0430 00:38:19.467737 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.467756 kubelet[3195]: W0430 00:38:19.467750 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.467851 kubelet[3195]: E0430 00:38:19.467830 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.467969 kubelet[3195]: E0430 00:38:19.467953 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.467969 kubelet[3195]: W0430 00:38:19.467966 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.468120 kubelet[3195]: E0430 00:38:19.468030 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.468222 kubelet[3195]: E0430 00:38:19.468201 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.468222 kubelet[3195]: W0430 00:38:19.468214 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.468365 kubelet[3195]: E0430 00:38:19.468288 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.468448 kubelet[3195]: E0430 00:38:19.468429 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.468448 kubelet[3195]: W0430 00:38:19.468441 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.468508 kubelet[3195]: E0430 00:38:19.468457 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.468657 kubelet[3195]: E0430 00:38:19.468638 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.468657 kubelet[3195]: W0430 00:38:19.468653 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.468699 kubelet[3195]: E0430 00:38:19.468669 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.468873 kubelet[3195]: E0430 00:38:19.468853 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.468873 kubelet[3195]: W0430 00:38:19.468867 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.468937 kubelet[3195]: E0430 00:38:19.468877 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.469188 kubelet[3195]: E0430 00:38:19.469166 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.469188 kubelet[3195]: W0430 00:38:19.469183 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.469256 kubelet[3195]: E0430 00:38:19.469193 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.469434 kubelet[3195]: E0430 00:38:19.469413 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.469434 kubelet[3195]: W0430 00:38:19.469428 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.469498 kubelet[3195]: E0430 00:38:19.469437 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.469657 kubelet[3195]: E0430 00:38:19.469635 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.469657 kubelet[3195]: W0430 00:38:19.469653 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.469713 kubelet[3195]: E0430 00:38:19.469662 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.469840 kubelet[3195]: E0430 00:38:19.469822 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.469840 kubelet[3195]: W0430 00:38:19.469837 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.469890 kubelet[3195]: E0430 00:38:19.469847 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.505819 kubelet[3195]: E0430 00:38:19.505700 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.505819 kubelet[3195]: W0430 00:38:19.505747 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.505819 kubelet[3195]: E0430 00:38:19.505770 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.568877 kubelet[3195]: E0430 00:38:19.568850 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.569151 kubelet[3195]: W0430 00:38:19.569008 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.569151 kubelet[3195]: E0430 00:38:19.569034 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.569571 kubelet[3195]: E0430 00:38:19.569428 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.569571 kubelet[3195]: W0430 00:38:19.569443 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.569571 kubelet[3195]: E0430 00:38:19.569465 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.569885 kubelet[3195]: E0430 00:38:19.569772 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.569885 kubelet[3195]: W0430 00:38:19.569790 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.569885 kubelet[3195]: E0430 00:38:19.569823 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.570271 kubelet[3195]: E0430 00:38:19.570186 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.570271 kubelet[3195]: W0430 00:38:19.570200 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.570271 kubelet[3195]: E0430 00:38:19.570221 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.570517 kubelet[3195]: E0430 00:38:19.570490 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.570517 kubelet[3195]: W0430 00:38:19.570509 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.570517 kubelet[3195]: E0430 00:38:19.570533 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.570707 kubelet[3195]: E0430 00:38:19.570692 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.570707 kubelet[3195]: W0430 00:38:19.570704 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.570840 kubelet[3195]: E0430 00:38:19.570721 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.571023 kubelet[3195]: E0430 00:38:19.571008 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.571023 kubelet[3195]: W0430 00:38:19.571021 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.571411 kubelet[3195]: E0430 00:38:19.571052 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.571411 kubelet[3195]: E0430 00:38:19.571188 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.571411 kubelet[3195]: W0430 00:38:19.571301 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.571574 kubelet[3195]: E0430 00:38:19.571546 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.571698 kubelet[3195]: E0430 00:38:19.571678 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.571810 kubelet[3195]: W0430 00:38:19.571751 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.571810 kubelet[3195]: E0430 00:38:19.571798 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.572179 kubelet[3195]: E0430 00:38:19.572105 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.572179 kubelet[3195]: W0430 00:38:19.572119 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.572179 kubelet[3195]: E0430 00:38:19.572151 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.572615 kubelet[3195]: E0430 00:38:19.572550 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.572615 kubelet[3195]: W0430 00:38:19.572566 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.572615 kubelet[3195]: E0430 00:38:19.572605 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.573934 kubelet[3195]: E0430 00:38:19.573825 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.573934 kubelet[3195]: W0430 00:38:19.573849 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.573934 kubelet[3195]: E0430 00:38:19.573910 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.574868 kubelet[3195]: E0430 00:38:19.574837 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.575253 kubelet[3195]: W0430 00:38:19.575147 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.575546 kubelet[3195]: E0430 00:38:19.575348 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.575964 kubelet[3195]: E0430 00:38:19.575941 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.576403 kubelet[3195]: W0430 00:38:19.576157 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.576403 kubelet[3195]: E0430 00:38:19.576250 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.577273 kubelet[3195]: E0430 00:38:19.577248 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.577510 kubelet[3195]: W0430 00:38:19.577367 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.577510 kubelet[3195]: E0430 00:38:19.577438 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.577746 kubelet[3195]: E0430 00:38:19.577730 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.577808 kubelet[3195]: W0430 00:38:19.577796 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.577910 kubelet[3195]: E0430 00:38:19.577873 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.578765 kubelet[3195]: E0430 00:38:19.578634 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.578765 kubelet[3195]: W0430 00:38:19.578657 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.578854 kubelet[3195]: E0430 00:38:19.578718 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.579130 kubelet[3195]: E0430 00:38:19.579012 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.579130 kubelet[3195]: W0430 00:38:19.579025 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.579130 kubelet[3195]: E0430 00:38:19.579061 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.579573 kubelet[3195]: E0430 00:38:19.579366 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.579573 kubelet[3195]: W0430 00:38:19.579382 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.579573 kubelet[3195]: E0430 00:38:19.579441 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.579907 kubelet[3195]: E0430 00:38:19.579762 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.579907 kubelet[3195]: W0430 00:38:19.579778 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.579907 kubelet[3195]: E0430 00:38:19.579813 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.580171 kubelet[3195]: E0430 00:38:19.580155 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.580248 kubelet[3195]: W0430 00:38:19.580236 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.580439 kubelet[3195]: E0430 00:38:19.580309 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.580645 kubelet[3195]: E0430 00:38:19.580611 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.580717 kubelet[3195]: W0430 00:38:19.580705 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.580901 kubelet[3195]: E0430 00:38:19.580772 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.581172 kubelet[3195]: E0430 00:38:19.581154 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.581371 kubelet[3195]: W0430 00:38:19.581227 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.581371 kubelet[3195]: E0430 00:38:19.581253 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.581708 kubelet[3195]: E0430 00:38:19.581690 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.581875 kubelet[3195]: W0430 00:38:19.581768 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.581875 kubelet[3195]: E0430 00:38:19.581812 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.582158 kubelet[3195]: E0430 00:38:19.582104 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.582158 kubelet[3195]: W0430 00:38:19.582118 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.582158 kubelet[3195]: E0430 00:38:19.582130 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.649154 kubelet[3195]: E0430 00:38:19.648585 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:19.649154 kubelet[3195]: W0430 00:38:19.648615 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:19.649154 kubelet[3195]: E0430 00:38:19.648642 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:19.653746 containerd[1704]: time="2025-04-30T00:38:19.652620204Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-9jgff,Uid:7570c5da-24ed-4e7c-9b5f-ed2619cd28d9,Namespace:calico-system,Attempt:0,}" Apr 30 00:38:19.726720 containerd[1704]: time="2025-04-30T00:38:19.726552632Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:38:19.726720 containerd[1704]: time="2025-04-30T00:38:19.726624672Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:38:19.726720 containerd[1704]: time="2025-04-30T00:38:19.726641272Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:19.727233 containerd[1704]: time="2025-04-30T00:38:19.726770152Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:19.745273 systemd[1]: Started cri-containerd-90bb9a1599b9e47c9d36951ddbcd3aae8700f36cb9db9f2e7ba366fe64631460.scope - libcontainer container 90bb9a1599b9e47c9d36951ddbcd3aae8700f36cb9db9f2e7ba366fe64631460. Apr 30 00:38:19.769786 containerd[1704]: time="2025-04-30T00:38:19.769695391Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-9jgff,Uid:7570c5da-24ed-4e7c-9b5f-ed2619cd28d9,Namespace:calico-system,Attempt:0,} returns sandbox id \"90bb9a1599b9e47c9d36951ddbcd3aae8700f36cb9db9f2e7ba366fe64631460\"" Apr 30 00:38:21.088949 containerd[1704]: time="2025-04-30T00:38:21.088885867Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:21.094202 containerd[1704]: time="2025-04-30T00:38:21.094150152Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.3: active requests=0, bytes read=28370571" Apr 30 00:38:21.103244 containerd[1704]: time="2025-04-30T00:38:21.102145519Z" level=info msg="ImageCreate event name:\"sha256:26e730979a07ea7452715da6ac48076016018bc982c06ebd32d5e095f42d3d54\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:21.116805 containerd[1704]: time="2025-04-30T00:38:21.116739933Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:21.117749 containerd[1704]: time="2025-04-30T00:38:21.117705494Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.3\" with image id \"sha256:26e730979a07ea7452715da6ac48076016018bc982c06ebd32d5e095f42d3d54\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\", size \"29739745\" in 1.904776498s" Apr 30 00:38:21.117891 containerd[1704]: time="2025-04-30T00:38:21.117874974Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\" returns image reference \"sha256:26e730979a07ea7452715da6ac48076016018bc982c06ebd32d5e095f42d3d54\"" Apr 30 00:38:21.119705 containerd[1704]: time="2025-04-30T00:38:21.119325055Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\"" Apr 30 00:38:21.133476 containerd[1704]: time="2025-04-30T00:38:21.133421428Z" level=info msg="CreateContainer within sandbox \"b0628cd2226d9ea1ec6b12cbae1d3c0d5097b82551c22a00cfd43bd30ea6b7b5\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Apr 30 00:38:21.177285 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount537454284.mount: Deactivated successfully. Apr 30 00:38:21.194740 containerd[1704]: time="2025-04-30T00:38:21.194628603Z" level=info msg="CreateContainer within sandbox \"b0628cd2226d9ea1ec6b12cbae1d3c0d5097b82551c22a00cfd43bd30ea6b7b5\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"6de9fdf88331669931fd1afa93f423a2daeb103c00f21b1c1ee8b7485fa2e93d\"" Apr 30 00:38:21.196270 containerd[1704]: time="2025-04-30T00:38:21.195595164Z" level=info msg="StartContainer for \"6de9fdf88331669931fd1afa93f423a2daeb103c00f21b1c1ee8b7485fa2e93d\"" Apr 30 00:38:21.229292 systemd[1]: Started cri-containerd-6de9fdf88331669931fd1afa93f423a2daeb103c00f21b1c1ee8b7485fa2e93d.scope - libcontainer container 6de9fdf88331669931fd1afa93f423a2daeb103c00f21b1c1ee8b7485fa2e93d. Apr 30 00:38:21.268015 containerd[1704]: time="2025-04-30T00:38:21.267932870Z" level=info msg="StartContainer for \"6de9fdf88331669931fd1afa93f423a2daeb103c00f21b1c1ee8b7485fa2e93d\" returns successfully" Apr 30 00:38:21.275833 kubelet[3195]: E0430 00:38:21.275756 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:21.414706 kubelet[3195]: I0430 00:38:21.414539 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-5dd8879757-zqhv7" podStartSLOduration=1.508181783 podStartE2EDuration="3.414506683s" podCreationTimestamp="2025-04-30 00:38:18 +0000 UTC" firstStartedPulling="2025-04-30 00:38:19.212494235 +0000 UTC m=+14.047720403" lastFinishedPulling="2025-04-30 00:38:21.118819175 +0000 UTC m=+15.954045303" observedRunningTime="2025-04-30 00:38:21.412612401 +0000 UTC m=+16.247838569" watchObservedRunningTime="2025-04-30 00:38:21.414506683 +0000 UTC m=+16.249732851" Apr 30 00:38:21.481562 kubelet[3195]: E0430 00:38:21.481527 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.481562 kubelet[3195]: W0430 00:38:21.481553 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.481562 kubelet[3195]: E0430 00:38:21.481575 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.481803 kubelet[3195]: E0430 00:38:21.481780 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.481803 kubelet[3195]: W0430 00:38:21.481799 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.481881 kubelet[3195]: E0430 00:38:21.481809 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.481986 kubelet[3195]: E0430 00:38:21.481972 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.482021 kubelet[3195]: W0430 00:38:21.481994 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.482021 kubelet[3195]: E0430 00:38:21.482004 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.482274 kubelet[3195]: E0430 00:38:21.482256 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.482274 kubelet[3195]: W0430 00:38:21.482269 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.482363 kubelet[3195]: E0430 00:38:21.482286 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.482504 kubelet[3195]: E0430 00:38:21.482486 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.482504 kubelet[3195]: W0430 00:38:21.482501 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.482552 kubelet[3195]: E0430 00:38:21.482510 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.482685 kubelet[3195]: E0430 00:38:21.482670 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.482685 kubelet[3195]: W0430 00:38:21.482681 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.482764 kubelet[3195]: E0430 00:38:21.482700 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.482857 kubelet[3195]: E0430 00:38:21.482841 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.482895 kubelet[3195]: W0430 00:38:21.482862 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.482895 kubelet[3195]: E0430 00:38:21.482871 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.483038 kubelet[3195]: E0430 00:38:21.483020 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.483038 kubelet[3195]: W0430 00:38:21.483033 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.483038 kubelet[3195]: E0430 00:38:21.483041 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.483261 kubelet[3195]: E0430 00:38:21.483235 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.483261 kubelet[3195]: W0430 00:38:21.483257 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.483343 kubelet[3195]: E0430 00:38:21.483266 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.483471 kubelet[3195]: E0430 00:38:21.483447 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.483471 kubelet[3195]: W0430 00:38:21.483459 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.483471 kubelet[3195]: E0430 00:38:21.483467 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.483670 kubelet[3195]: E0430 00:38:21.483616 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.483670 kubelet[3195]: W0430 00:38:21.483623 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.483670 kubelet[3195]: E0430 00:38:21.483630 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.483925 kubelet[3195]: E0430 00:38:21.483772 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.483925 kubelet[3195]: W0430 00:38:21.483784 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.483925 kubelet[3195]: E0430 00:38:21.483793 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.484332 kubelet[3195]: E0430 00:38:21.484301 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.484509 kubelet[3195]: W0430 00:38:21.484424 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.484509 kubelet[3195]: E0430 00:38:21.484444 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.484820 kubelet[3195]: E0430 00:38:21.484801 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.485008 kubelet[3195]: W0430 00:38:21.484893 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.485008 kubelet[3195]: E0430 00:38:21.484911 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.485211 kubelet[3195]: E0430 00:38:21.485196 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.485267 kubelet[3195]: W0430 00:38:21.485256 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.485328 kubelet[3195]: E0430 00:38:21.485318 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.485791 kubelet[3195]: E0430 00:38:21.485762 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.485791 kubelet[3195]: W0430 00:38:21.485784 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.485875 kubelet[3195]: E0430 00:38:21.485796 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.486046 kubelet[3195]: E0430 00:38:21.486023 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.486046 kubelet[3195]: W0430 00:38:21.486039 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.486215 kubelet[3195]: E0430 00:38:21.486059 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.486319 kubelet[3195]: E0430 00:38:21.486295 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.486319 kubelet[3195]: W0430 00:38:21.486311 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.486395 kubelet[3195]: E0430 00:38:21.486328 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.486627 kubelet[3195]: E0430 00:38:21.486605 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.486627 kubelet[3195]: W0430 00:38:21.486621 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.486706 kubelet[3195]: E0430 00:38:21.486648 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.486872 kubelet[3195]: E0430 00:38:21.486834 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.486872 kubelet[3195]: W0430 00:38:21.486852 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.486872 kubelet[3195]: E0430 00:38:21.486869 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.487198 kubelet[3195]: E0430 00:38:21.487177 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.487198 kubelet[3195]: W0430 00:38:21.487193 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.487279 kubelet[3195]: E0430 00:38:21.487217 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.487743 kubelet[3195]: E0430 00:38:21.487717 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.487743 kubelet[3195]: W0430 00:38:21.487738 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.487850 kubelet[3195]: E0430 00:38:21.487757 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.488109 kubelet[3195]: E0430 00:38:21.488087 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.488109 kubelet[3195]: W0430 00:38:21.488104 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.488263 kubelet[3195]: E0430 00:38:21.488218 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.488335 kubelet[3195]: E0430 00:38:21.488318 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.488335 kubelet[3195]: W0430 00:38:21.488331 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.488420 kubelet[3195]: E0430 00:38:21.488399 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.488543 kubelet[3195]: E0430 00:38:21.488521 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.488543 kubelet[3195]: W0430 00:38:21.488536 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.488684 kubelet[3195]: E0430 00:38:21.488550 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.488765 kubelet[3195]: E0430 00:38:21.488721 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.488765 kubelet[3195]: W0430 00:38:21.488741 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.488765 kubelet[3195]: E0430 00:38:21.488750 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.488948 kubelet[3195]: E0430 00:38:21.488876 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.488948 kubelet[3195]: W0430 00:38:21.488882 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.488948 kubelet[3195]: E0430 00:38:21.488899 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.489120 kubelet[3195]: E0430 00:38:21.489101 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.489120 kubelet[3195]: W0430 00:38:21.489116 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.489180 kubelet[3195]: E0430 00:38:21.489133 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.489481 kubelet[3195]: E0430 00:38:21.489460 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.489481 kubelet[3195]: W0430 00:38:21.489478 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.489554 kubelet[3195]: E0430 00:38:21.489493 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.489670 kubelet[3195]: E0430 00:38:21.489647 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.489670 kubelet[3195]: W0430 00:38:21.489664 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.489731 kubelet[3195]: E0430 00:38:21.489685 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.489926 kubelet[3195]: E0430 00:38:21.489909 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.489926 kubelet[3195]: W0430 00:38:21.489923 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.489990 kubelet[3195]: E0430 00:38:21.489942 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.490329 kubelet[3195]: E0430 00:38:21.490303 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.490329 kubelet[3195]: W0430 00:38:21.490325 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.490424 kubelet[3195]: E0430 00:38:21.490342 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:21.490552 kubelet[3195]: E0430 00:38:21.490533 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:21.490552 kubelet[3195]: W0430 00:38:21.490550 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:21.490615 kubelet[3195]: E0430 00:38:21.490563 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.366735 containerd[1704]: time="2025-04-30T00:38:22.366658307Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:22.370517 containerd[1704]: time="2025-04-30T00:38:22.370465270Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3: active requests=0, bytes read=5122903" Apr 30 00:38:22.378717 containerd[1704]: time="2025-04-30T00:38:22.378634837Z" level=info msg="ImageCreate event name:\"sha256:dd8e710a588cc6f5834c4d84f7e12458efae593d3dfe527ca9e757c89239ecb8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:22.383595 containerd[1704]: time="2025-04-30T00:38:22.383521482Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:22.384446 containerd[1704]: time="2025-04-30T00:38:22.384286643Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" with image id \"sha256:dd8e710a588cc6f5834c4d84f7e12458efae593d3dfe527ca9e757c89239ecb8\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\", size \"6492045\" in 1.264919308s" Apr 30 00:38:22.384446 containerd[1704]: time="2025-04-30T00:38:22.384332603Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" returns image reference \"sha256:dd8e710a588cc6f5834c4d84f7e12458efae593d3dfe527ca9e757c89239ecb8\"" Apr 30 00:38:22.386927 containerd[1704]: time="2025-04-30T00:38:22.386736565Z" level=info msg="CreateContainer within sandbox \"90bb9a1599b9e47c9d36951ddbcd3aae8700f36cb9db9f2e7ba366fe64631460\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Apr 30 00:38:22.391499 kubelet[3195]: I0430 00:38:22.391099 3195 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 00:38:22.392103 kubelet[3195]: E0430 00:38:22.391959 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.392103 kubelet[3195]: W0430 00:38:22.391978 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.392103 kubelet[3195]: E0430 00:38:22.391996 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.392568 kubelet[3195]: E0430 00:38:22.392178 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.392568 kubelet[3195]: W0430 00:38:22.392188 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.392568 kubelet[3195]: E0430 00:38:22.392196 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.392568 kubelet[3195]: E0430 00:38:22.392322 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.392568 kubelet[3195]: W0430 00:38:22.392329 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.392568 kubelet[3195]: E0430 00:38:22.392343 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.392568 kubelet[3195]: E0430 00:38:22.392480 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.392568 kubelet[3195]: W0430 00:38:22.392488 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.392568 kubelet[3195]: E0430 00:38:22.392496 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.392771 kubelet[3195]: E0430 00:38:22.392635 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.392771 kubelet[3195]: W0430 00:38:22.392643 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.392771 kubelet[3195]: E0430 00:38:22.392651 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.392771 kubelet[3195]: E0430 00:38:22.392770 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.392850 kubelet[3195]: W0430 00:38:22.392776 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.392850 kubelet[3195]: E0430 00:38:22.392787 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.392998 kubelet[3195]: E0430 00:38:22.392898 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.392998 kubelet[3195]: W0430 00:38:22.392913 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.392998 kubelet[3195]: E0430 00:38:22.392921 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.393114 kubelet[3195]: E0430 00:38:22.393049 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.393114 kubelet[3195]: W0430 00:38:22.393056 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.393114 kubelet[3195]: E0430 00:38:22.393063 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.393255 kubelet[3195]: E0430 00:38:22.393239 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.393255 kubelet[3195]: W0430 00:38:22.393253 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.393318 kubelet[3195]: E0430 00:38:22.393263 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.393415 kubelet[3195]: E0430 00:38:22.393401 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.393415 kubelet[3195]: W0430 00:38:22.393413 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.393481 kubelet[3195]: E0430 00:38:22.393421 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.393596 kubelet[3195]: E0430 00:38:22.393581 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.393596 kubelet[3195]: W0430 00:38:22.393594 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.393653 kubelet[3195]: E0430 00:38:22.393603 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.393747 kubelet[3195]: E0430 00:38:22.393729 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.393747 kubelet[3195]: W0430 00:38:22.393741 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.393817 kubelet[3195]: E0430 00:38:22.393749 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.393982 kubelet[3195]: E0430 00:38:22.393962 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.393982 kubelet[3195]: W0430 00:38:22.393978 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.394039 kubelet[3195]: E0430 00:38:22.393988 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.394188 kubelet[3195]: E0430 00:38:22.394169 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.394188 kubelet[3195]: W0430 00:38:22.394185 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.394253 kubelet[3195]: E0430 00:38:22.394194 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.394356 kubelet[3195]: E0430 00:38:22.394335 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.394356 kubelet[3195]: W0430 00:38:22.394354 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.394416 kubelet[3195]: E0430 00:38:22.394363 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.394664 kubelet[3195]: E0430 00:38:22.394635 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.394664 kubelet[3195]: W0430 00:38:22.394657 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.394664 kubelet[3195]: E0430 00:38:22.394667 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.394877 kubelet[3195]: E0430 00:38:22.394859 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.394877 kubelet[3195]: W0430 00:38:22.394874 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.394938 kubelet[3195]: E0430 00:38:22.394894 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.395184 kubelet[3195]: E0430 00:38:22.395163 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.395184 kubelet[3195]: W0430 00:38:22.395179 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.395271 kubelet[3195]: E0430 00:38:22.395196 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.395441 kubelet[3195]: E0430 00:38:22.395400 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.395441 kubelet[3195]: W0430 00:38:22.395413 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.395441 kubelet[3195]: E0430 00:38:22.395432 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.395621 kubelet[3195]: E0430 00:38:22.395601 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.395621 kubelet[3195]: W0430 00:38:22.395618 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.395826 kubelet[3195]: E0430 00:38:22.395633 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.395925 kubelet[3195]: E0430 00:38:22.395905 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.395985 kubelet[3195]: W0430 00:38:22.395972 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.396049 kubelet[3195]: E0430 00:38:22.396037 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.396440 kubelet[3195]: E0430 00:38:22.396330 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.396440 kubelet[3195]: W0430 00:38:22.396346 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.396440 kubelet[3195]: E0430 00:38:22.396373 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.396626 kubelet[3195]: E0430 00:38:22.396612 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.396678 kubelet[3195]: W0430 00:38:22.396667 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.396799 kubelet[3195]: E0430 00:38:22.396763 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.397178 kubelet[3195]: E0430 00:38:22.397044 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.397178 kubelet[3195]: W0430 00:38:22.397060 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.397178 kubelet[3195]: E0430 00:38:22.397110 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.397620 kubelet[3195]: E0430 00:38:22.397483 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.397620 kubelet[3195]: W0430 00:38:22.397504 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.397620 kubelet[3195]: E0430 00:38:22.397527 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.398203 kubelet[3195]: E0430 00:38:22.397971 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.398203 kubelet[3195]: W0430 00:38:22.397988 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.398203 kubelet[3195]: E0430 00:38:22.398010 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.398932 kubelet[3195]: E0430 00:38:22.398912 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.399207 kubelet[3195]: W0430 00:38:22.399052 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.399207 kubelet[3195]: E0430 00:38:22.399109 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.399547 kubelet[3195]: E0430 00:38:22.399457 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.399547 kubelet[3195]: W0430 00:38:22.399470 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.399547 kubelet[3195]: E0430 00:38:22.399538 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.400685 kubelet[3195]: E0430 00:38:22.400058 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.400685 kubelet[3195]: W0430 00:38:22.400604 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.400778 kubelet[3195]: E0430 00:38:22.400744 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.401429 kubelet[3195]: E0430 00:38:22.401019 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.401429 kubelet[3195]: W0430 00:38:22.401035 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.401429 kubelet[3195]: E0430 00:38:22.401311 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.401570 kubelet[3195]: E0430 00:38:22.401486 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.401570 kubelet[3195]: W0430 00:38:22.401499 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.401570 kubelet[3195]: E0430 00:38:22.401517 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.402247 kubelet[3195]: E0430 00:38:22.402042 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.402247 kubelet[3195]: W0430 00:38:22.402109 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.402247 kubelet[3195]: E0430 00:38:22.402136 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.403519 kubelet[3195]: E0430 00:38:22.403012 3195 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 00:38:22.403519 kubelet[3195]: W0430 00:38:22.403037 3195 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 00:38:22.403519 kubelet[3195]: E0430 00:38:22.403054 3195 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 00:38:22.442505 containerd[1704]: time="2025-04-30T00:38:22.442428455Z" level=info msg="CreateContainer within sandbox \"90bb9a1599b9e47c9d36951ddbcd3aae8700f36cb9db9f2e7ba366fe64631460\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"63ebc816eeb7a40c8608305f84f279ed7346de5821ba7854357a4262417be5a0\"" Apr 30 00:38:22.443712 containerd[1704]: time="2025-04-30T00:38:22.443645576Z" level=info msg="StartContainer for \"63ebc816eeb7a40c8608305f84f279ed7346de5821ba7854357a4262417be5a0\"" Apr 30 00:38:22.491267 systemd[1]: Started cri-containerd-63ebc816eeb7a40c8608305f84f279ed7346de5821ba7854357a4262417be5a0.scope - libcontainer container 63ebc816eeb7a40c8608305f84f279ed7346de5821ba7854357a4262417be5a0. Apr 30 00:38:22.538355 containerd[1704]: time="2025-04-30T00:38:22.538238222Z" level=info msg="StartContainer for \"63ebc816eeb7a40c8608305f84f279ed7346de5821ba7854357a4262417be5a0\" returns successfully" Apr 30 00:38:22.551460 systemd[1]: cri-containerd-63ebc816eeb7a40c8608305f84f279ed7346de5821ba7854357a4262417be5a0.scope: Deactivated successfully. Apr 30 00:38:22.573005 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-63ebc816eeb7a40c8608305f84f279ed7346de5821ba7854357a4262417be5a0-rootfs.mount: Deactivated successfully. Apr 30 00:38:23.274955 kubelet[3195]: E0430 00:38:23.274425 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:23.494396 containerd[1704]: time="2025-04-30T00:38:23.494320089Z" level=info msg="shim disconnected" id=63ebc816eeb7a40c8608305f84f279ed7346de5821ba7854357a4262417be5a0 namespace=k8s.io Apr 30 00:38:23.494850 containerd[1704]: time="2025-04-30T00:38:23.494457450Z" level=warning msg="cleaning up after shim disconnected" id=63ebc816eeb7a40c8608305f84f279ed7346de5821ba7854357a4262417be5a0 namespace=k8s.io Apr 30 00:38:23.494850 containerd[1704]: time="2025-04-30T00:38:23.494468770Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 30 00:38:24.400036 containerd[1704]: time="2025-04-30T00:38:24.399982791Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\"" Apr 30 00:38:25.275267 kubelet[3195]: E0430 00:38:25.274626 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:27.275881 kubelet[3195]: E0430 00:38:27.274526 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:29.276103 kubelet[3195]: E0430 00:38:29.274838 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:31.275811 kubelet[3195]: E0430 00:38:31.274594 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:31.718233 containerd[1704]: time="2025-04-30T00:38:31.718174665Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:31.723079 containerd[1704]: time="2025-04-30T00:38:31.723009109Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.3: active requests=0, bytes read=91256270" Apr 30 00:38:31.766617 containerd[1704]: time="2025-04-30T00:38:31.766561510Z" level=info msg="ImageCreate event name:\"sha256:add6372545fb406bb017769f222d84c50549ce13e3b19f1fbaee3d8a4aaef627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:31.811952 containerd[1704]: time="2025-04-30T00:38:31.811872832Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:31.812786 containerd[1704]: time="2025-04-30T00:38:31.812676193Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.3\" with image id \"sha256:add6372545fb406bb017769f222d84c50549ce13e3b19f1fbaee3d8a4aaef627\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\", size \"92625452\" in 7.412625522s" Apr 30 00:38:31.812786 containerd[1704]: time="2025-04-30T00:38:31.812710153Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\" returns image reference \"sha256:add6372545fb406bb017769f222d84c50549ce13e3b19f1fbaee3d8a4aaef627\"" Apr 30 00:38:31.816218 containerd[1704]: time="2025-04-30T00:38:31.816156836Z" level=info msg="CreateContainer within sandbox \"90bb9a1599b9e47c9d36951ddbcd3aae8700f36cb9db9f2e7ba366fe64631460\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Apr 30 00:38:32.117362 containerd[1704]: time="2025-04-30T00:38:32.117237317Z" level=info msg="CreateContainer within sandbox \"90bb9a1599b9e47c9d36951ddbcd3aae8700f36cb9db9f2e7ba366fe64631460\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"35bb3f0f8c4921e9e9dc26792c3df304178b14536229446d96229b2313138b76\"" Apr 30 00:38:32.118369 containerd[1704]: time="2025-04-30T00:38:32.118079037Z" level=info msg="StartContainer for \"35bb3f0f8c4921e9e9dc26792c3df304178b14536229446d96229b2313138b76\"" Apr 30 00:38:32.151319 systemd[1]: Started cri-containerd-35bb3f0f8c4921e9e9dc26792c3df304178b14536229446d96229b2313138b76.scope - libcontainer container 35bb3f0f8c4921e9e9dc26792c3df304178b14536229446d96229b2313138b76. Apr 30 00:38:32.189493 containerd[1704]: time="2025-04-30T00:38:32.189355704Z" level=info msg="StartContainer for \"35bb3f0f8c4921e9e9dc26792c3df304178b14536229446d96229b2313138b76\" returns successfully" Apr 30 00:38:33.276016 kubelet[3195]: E0430 00:38:33.274716 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:35.275456 kubelet[3195]: E0430 00:38:35.275352 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:37.277391 kubelet[3195]: E0430 00:38:37.276710 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:39.017475 kubelet[3195]: I0430 00:38:39.017223 3195 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 00:38:43.218446 kubelet[3195]: E0430 00:38:39.274690 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:43.218446 kubelet[3195]: E0430 00:38:41.274287 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:43.275687 kubelet[3195]: E0430 00:38:43.275300 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:43.892564 containerd[1704]: time="2025-04-30T00:38:43.892509069Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 30 00:38:43.894971 systemd[1]: cri-containerd-35bb3f0f8c4921e9e9dc26792c3df304178b14536229446d96229b2313138b76.scope: Deactivated successfully. Apr 30 00:38:43.906284 kubelet[3195]: I0430 00:38:43.905816 3195 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Apr 30 00:38:43.918469 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-35bb3f0f8c4921e9e9dc26792c3df304178b14536229446d96229b2313138b76-rootfs.mount: Deactivated successfully. Apr 30 00:38:43.935509 containerd[1704]: time="2025-04-30T00:38:43.935415312Z" level=info msg="shim disconnected" id=35bb3f0f8c4921e9e9dc26792c3df304178b14536229446d96229b2313138b76 namespace=k8s.io Apr 30 00:38:43.935509 containerd[1704]: time="2025-04-30T00:38:43.935488952Z" level=warning msg="cleaning up after shim disconnected" id=35bb3f0f8c4921e9e9dc26792c3df304178b14536229446d96229b2313138b76 namespace=k8s.io Apr 30 00:38:43.935509 containerd[1704]: time="2025-04-30T00:38:43.935499912Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 30 00:38:43.997157 systemd[1]: Created slice kubepods-burstable-podf93f7ee2_841a_4f1b_a68e_709e05dec366.slice - libcontainer container kubepods-burstable-podf93f7ee2_841a_4f1b_a68e_709e05dec366.slice. Apr 30 00:38:43.999547 kubelet[3195]: W0430 00:38:43.998794 3195 reflector.go:561] object-"calico-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4081.3.3-a-c388cdac8a" cannot list resource "configmaps" in API group "" in the namespace "calico-apiserver": no relationship found between node 'ci-4081.3.3-a-c388cdac8a' and this object Apr 30 00:38:43.999547 kubelet[3195]: E0430 00:38:43.998850 3195 reflector.go:158] "Unhandled Error" err="object-\"calico-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:ci-4081.3.3-a-c388cdac8a\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"calico-apiserver\": no relationship found between node 'ci-4081.3.3-a-c388cdac8a' and this object" logger="UnhandledError" Apr 30 00:38:43.999547 kubelet[3195]: W0430 00:38:43.998905 3195 reflector.go:561] object-"calico-apiserver"/"calico-apiserver-certs": failed to list *v1.Secret: secrets "calico-apiserver-certs" is forbidden: User "system:node:ci-4081.3.3-a-c388cdac8a" cannot list resource "secrets" in API group "" in the namespace "calico-apiserver": no relationship found between node 'ci-4081.3.3-a-c388cdac8a' and this object Apr 30 00:38:43.999547 kubelet[3195]: E0430 00:38:43.998916 3195 reflector.go:158] "Unhandled Error" err="object-\"calico-apiserver\"/\"calico-apiserver-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"calico-apiserver-certs\" is forbidden: User \"system:node:ci-4081.3.3-a-c388cdac8a\" cannot list resource \"secrets\" in API group \"\" in the namespace \"calico-apiserver\": no relationship found between node 'ci-4081.3.3-a-c388cdac8a' and this object" logger="UnhandledError" Apr 30 00:38:44.007382 systemd[1]: Created slice kubepods-burstable-pod2d0e9630_4ffc_4065_9430_75363f7abec1.slice - libcontainer container kubepods-burstable-pod2d0e9630_4ffc_4065_9430_75363f7abec1.slice. Apr 30 00:38:44.013218 systemd[1]: Created slice kubepods-besteffort-pod8f15ee70_a765_40e4_8bb0_3123a84281b8.slice - libcontainer container kubepods-besteffort-pod8f15ee70_a765_40e4_8bb0_3123a84281b8.slice. Apr 30 00:38:44.019163 systemd[1]: Created slice kubepods-besteffort-pod1be863bb_f182_4a93_978b_d215d3fc0c5d.slice - libcontainer container kubepods-besteffort-pod1be863bb_f182_4a93_978b_d215d3fc0c5d.slice. Apr 30 00:38:44.024784 systemd[1]: Created slice kubepods-besteffort-podf43c40a3_2edd_4744_8016_2953f9124512.slice - libcontainer container kubepods-besteffort-podf43c40a3_2edd_4744_8016_2953f9124512.slice. Apr 30 00:38:44.143884 kubelet[3195]: I0430 00:38:44.143702 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d0e9630-4ffc-4065-9430-75363f7abec1-config-volume\") pod \"coredns-6f6b679f8f-bgzm6\" (UID: \"2d0e9630-4ffc-4065-9430-75363f7abec1\") " pod="kube-system/coredns-6f6b679f8f-bgzm6" Apr 30 00:38:44.143884 kubelet[3195]: I0430 00:38:44.143748 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd5fr\" (UniqueName: \"kubernetes.io/projected/f93f7ee2-841a-4f1b-a68e-709e05dec366-kube-api-access-qd5fr\") pod \"coredns-6f6b679f8f-7z7rd\" (UID: \"f93f7ee2-841a-4f1b-a68e-709e05dec366\") " pod="kube-system/coredns-6f6b679f8f-7z7rd" Apr 30 00:38:44.143884 kubelet[3195]: I0430 00:38:44.143767 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkmn8\" (UniqueName: \"kubernetes.io/projected/8f15ee70-a765-40e4-8bb0-3123a84281b8-kube-api-access-dkmn8\") pod \"calico-kube-controllers-768b5d7cdb-g4mjx\" (UID: \"8f15ee70-a765-40e4-8bb0-3123a84281b8\") " pod="calico-system/calico-kube-controllers-768b5d7cdb-g4mjx" Apr 30 00:38:44.143884 kubelet[3195]: I0430 00:38:44.143793 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/f43c40a3-2edd-4744-8016-2953f9124512-calico-apiserver-certs\") pod \"calico-apiserver-7dd664d894-tvxc5\" (UID: \"f43c40a3-2edd-4744-8016-2953f9124512\") " pod="calico-apiserver/calico-apiserver-7dd664d894-tvxc5" Apr 30 00:38:44.143884 kubelet[3195]: I0430 00:38:44.143810 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f15ee70-a765-40e4-8bb0-3123a84281b8-tigera-ca-bundle\") pod \"calico-kube-controllers-768b5d7cdb-g4mjx\" (UID: \"8f15ee70-a765-40e4-8bb0-3123a84281b8\") " pod="calico-system/calico-kube-controllers-768b5d7cdb-g4mjx" Apr 30 00:38:44.144333 kubelet[3195]: I0430 00:38:44.143828 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f93f7ee2-841a-4f1b-a68e-709e05dec366-config-volume\") pod \"coredns-6f6b679f8f-7z7rd\" (UID: \"f93f7ee2-841a-4f1b-a68e-709e05dec366\") " pod="kube-system/coredns-6f6b679f8f-7z7rd" Apr 30 00:38:44.144333 kubelet[3195]: I0430 00:38:44.143846 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/1be863bb-f182-4a93-978b-d215d3fc0c5d-calico-apiserver-certs\") pod \"calico-apiserver-7dd664d894-lcc4f\" (UID: \"1be863bb-f182-4a93-978b-d215d3fc0c5d\") " pod="calico-apiserver/calico-apiserver-7dd664d894-lcc4f" Apr 30 00:38:44.144333 kubelet[3195]: I0430 00:38:44.143866 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2rpp\" (UniqueName: \"kubernetes.io/projected/1be863bb-f182-4a93-978b-d215d3fc0c5d-kube-api-access-z2rpp\") pod \"calico-apiserver-7dd664d894-lcc4f\" (UID: \"1be863bb-f182-4a93-978b-d215d3fc0c5d\") " pod="calico-apiserver/calico-apiserver-7dd664d894-lcc4f" Apr 30 00:38:44.144333 kubelet[3195]: I0430 00:38:44.143897 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q297l\" (UniqueName: \"kubernetes.io/projected/f43c40a3-2edd-4744-8016-2953f9124512-kube-api-access-q297l\") pod \"calico-apiserver-7dd664d894-tvxc5\" (UID: \"f43c40a3-2edd-4744-8016-2953f9124512\") " pod="calico-apiserver/calico-apiserver-7dd664d894-tvxc5" Apr 30 00:38:44.144333 kubelet[3195]: I0430 00:38:44.143933 3195 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q7w9\" (UniqueName: \"kubernetes.io/projected/2d0e9630-4ffc-4065-9430-75363f7abec1-kube-api-access-8q7w9\") pod \"coredns-6f6b679f8f-bgzm6\" (UID: \"2d0e9630-4ffc-4065-9430-75363f7abec1\") " pod="kube-system/coredns-6f6b679f8f-bgzm6" Apr 30 00:38:44.443114 containerd[1704]: time="2025-04-30T00:38:44.442772624Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\"" Apr 30 00:38:44.604221 containerd[1704]: time="2025-04-30T00:38:44.604173667Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-7z7rd,Uid:f93f7ee2-841a-4f1b-a68e-709e05dec366,Namespace:kube-system,Attempt:0,}" Apr 30 00:38:44.611414 containerd[1704]: time="2025-04-30T00:38:44.611114954Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-bgzm6,Uid:2d0e9630-4ffc-4065-9430-75363f7abec1,Namespace:kube-system,Attempt:0,}" Apr 30 00:38:44.616745 containerd[1704]: time="2025-04-30T00:38:44.616452640Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-768b5d7cdb-g4mjx,Uid:8f15ee70-a765-40e4-8bb0-3123a84281b8,Namespace:calico-system,Attempt:0,}" Apr 30 00:38:44.797891 containerd[1704]: time="2025-04-30T00:38:44.797757023Z" level=error msg="Failed to destroy network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.798751 containerd[1704]: time="2025-04-30T00:38:44.798662264Z" level=error msg="encountered an error cleaning up failed sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.799183 containerd[1704]: time="2025-04-30T00:38:44.799134184Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-7z7rd,Uid:f93f7ee2-841a-4f1b-a68e-709e05dec366,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.799772 kubelet[3195]: E0430 00:38:44.799464 3195 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.799772 kubelet[3195]: E0430 00:38:44.799562 3195 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-7z7rd" Apr 30 00:38:44.799772 kubelet[3195]: E0430 00:38:44.799585 3195 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-7z7rd" Apr 30 00:38:44.800153 kubelet[3195]: E0430 00:38:44.799654 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-7z7rd_kube-system(f93f7ee2-841a-4f1b-a68e-709e05dec366)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-7z7rd_kube-system(f93f7ee2-841a-4f1b-a68e-709e05dec366)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-7z7rd" podUID="f93f7ee2-841a-4f1b-a68e-709e05dec366" Apr 30 00:38:44.840141 containerd[1704]: time="2025-04-30T00:38:44.839302865Z" level=error msg="Failed to destroy network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.840141 containerd[1704]: time="2025-04-30T00:38:44.840040785Z" level=error msg="encountered an error cleaning up failed sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.841235 containerd[1704]: time="2025-04-30T00:38:44.840185306Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-bgzm6,Uid:2d0e9630-4ffc-4065-9430-75363f7abec1,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.841282 kubelet[3195]: E0430 00:38:44.840415 3195 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.841282 kubelet[3195]: E0430 00:38:44.840469 3195 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-bgzm6" Apr 30 00:38:44.841282 kubelet[3195]: E0430 00:38:44.840490 3195 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-bgzm6" Apr 30 00:38:44.841364 kubelet[3195]: E0430 00:38:44.840539 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-bgzm6_kube-system(2d0e9630-4ffc-4065-9430-75363f7abec1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-bgzm6_kube-system(2d0e9630-4ffc-4065-9430-75363f7abec1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-bgzm6" podUID="2d0e9630-4ffc-4065-9430-75363f7abec1" Apr 30 00:38:44.843403 containerd[1704]: time="2025-04-30T00:38:44.843347429Z" level=error msg="Failed to destroy network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.843741 containerd[1704]: time="2025-04-30T00:38:44.843709029Z" level=error msg="encountered an error cleaning up failed sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.843804 containerd[1704]: time="2025-04-30T00:38:44.843773709Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-768b5d7cdb-g4mjx,Uid:8f15ee70-a765-40e4-8bb0-3123a84281b8,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.844051 kubelet[3195]: E0430 00:38:44.844014 3195 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:44.844138 kubelet[3195]: E0430 00:38:44.844113 3195 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-768b5d7cdb-g4mjx" Apr 30 00:38:44.844177 kubelet[3195]: E0430 00:38:44.844136 3195 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-768b5d7cdb-g4mjx" Apr 30 00:38:44.844201 kubelet[3195]: E0430 00:38:44.844182 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-768b5d7cdb-g4mjx_calico-system(8f15ee70-a765-40e4-8bb0-3123a84281b8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-768b5d7cdb-g4mjx_calico-system(8f15ee70-a765-40e4-8bb0-3123a84281b8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-768b5d7cdb-g4mjx" podUID="8f15ee70-a765-40e4-8bb0-3123a84281b8" Apr 30 00:38:44.919643 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581-shm.mount: Deactivated successfully. Apr 30 00:38:45.245781 kubelet[3195]: E0430 00:38:45.245407 3195 secret.go:188] Couldn't get secret calico-apiserver/calico-apiserver-certs: failed to sync secret cache: timed out waiting for the condition Apr 30 00:38:45.245781 kubelet[3195]: E0430 00:38:45.245507 3195 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1be863bb-f182-4a93-978b-d215d3fc0c5d-calico-apiserver-certs podName:1be863bb-f182-4a93-978b-d215d3fc0c5d nodeName:}" failed. No retries permitted until 2025-04-30 00:38:45.745483915 +0000 UTC m=+40.580710083 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "calico-apiserver-certs" (UniqueName: "kubernetes.io/secret/1be863bb-f182-4a93-978b-d215d3fc0c5d-calico-apiserver-certs") pod "calico-apiserver-7dd664d894-lcc4f" (UID: "1be863bb-f182-4a93-978b-d215d3fc0c5d") : failed to sync secret cache: timed out waiting for the condition Apr 30 00:38:45.246133 kubelet[3195]: E0430 00:38:45.246116 3195 secret.go:188] Couldn't get secret calico-apiserver/calico-apiserver-certs: failed to sync secret cache: timed out waiting for the condition Apr 30 00:38:45.246288 kubelet[3195]: E0430 00:38:45.246217 3195 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43c40a3-2edd-4744-8016-2953f9124512-calico-apiserver-certs podName:f43c40a3-2edd-4744-8016-2953f9124512 nodeName:}" failed. No retries permitted until 2025-04-30 00:38:45.746202555 +0000 UTC m=+40.581428683 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "calico-apiserver-certs" (UniqueName: "kubernetes.io/secret/f43c40a3-2edd-4744-8016-2953f9124512-calico-apiserver-certs") pod "calico-apiserver-7dd664d894-tvxc5" (UID: "f43c40a3-2edd-4744-8016-2953f9124512") : failed to sync secret cache: timed out waiting for the condition Apr 30 00:38:45.282688 systemd[1]: Created slice kubepods-besteffort-pod633a0f7e_ae5b_47ef_baa9_358c2fe3b117.slice - libcontainer container kubepods-besteffort-pod633a0f7e_ae5b_47ef_baa9_358c2fe3b117.slice. Apr 30 00:38:45.288380 containerd[1704]: time="2025-04-30T00:38:45.286023076Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-shvzf,Uid:633a0f7e-ae5b-47ef-baa9-358c2fe3b117,Namespace:calico-system,Attempt:0,}" Apr 30 00:38:45.339264 kubelet[3195]: E0430 00:38:45.339216 3195 projected.go:288] Couldn't get configMap calico-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Apr 30 00:38:45.339264 kubelet[3195]: E0430 00:38:45.339265 3195 projected.go:194] Error preparing data for projected volume kube-api-access-z2rpp for pod calico-apiserver/calico-apiserver-7dd664d894-lcc4f: failed to sync configmap cache: timed out waiting for the condition Apr 30 00:38:45.339446 kubelet[3195]: E0430 00:38:45.339342 3195 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1be863bb-f182-4a93-978b-d215d3fc0c5d-kube-api-access-z2rpp podName:1be863bb-f182-4a93-978b-d215d3fc0c5d nodeName:}" failed. No retries permitted until 2025-04-30 00:38:45.839323089 +0000 UTC m=+40.674549217 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-z2rpp" (UniqueName: "kubernetes.io/projected/1be863bb-f182-4a93-978b-d215d3fc0c5d-kube-api-access-z2rpp") pod "calico-apiserver-7dd664d894-lcc4f" (UID: "1be863bb-f182-4a93-978b-d215d3fc0c5d") : failed to sync configmap cache: timed out waiting for the condition Apr 30 00:38:45.353732 kubelet[3195]: E0430 00:38:45.353562 3195 projected.go:288] Couldn't get configMap calico-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Apr 30 00:38:45.353732 kubelet[3195]: E0430 00:38:45.353605 3195 projected.go:194] Error preparing data for projected volume kube-api-access-q297l for pod calico-apiserver/calico-apiserver-7dd664d894-tvxc5: failed to sync configmap cache: timed out waiting for the condition Apr 30 00:38:45.353732 kubelet[3195]: E0430 00:38:45.353658 3195 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f43c40a3-2edd-4744-8016-2953f9124512-kube-api-access-q297l podName:f43c40a3-2edd-4744-8016-2953f9124512 nodeName:}" failed. No retries permitted until 2025-04-30 00:38:45.853640464 +0000 UTC m=+40.688866632 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-q297l" (UniqueName: "kubernetes.io/projected/f43c40a3-2edd-4744-8016-2953f9124512-kube-api-access-q297l") pod "calico-apiserver-7dd664d894-tvxc5" (UID: "f43c40a3-2edd-4744-8016-2953f9124512") : failed to sync configmap cache: timed out waiting for the condition Apr 30 00:38:45.404953 containerd[1704]: time="2025-04-30T00:38:45.404877796Z" level=error msg="Failed to destroy network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:45.405354 containerd[1704]: time="2025-04-30T00:38:45.405320516Z" level=error msg="encountered an error cleaning up failed sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:45.405432 containerd[1704]: time="2025-04-30T00:38:45.405391596Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-shvzf,Uid:633a0f7e-ae5b-47ef-baa9-358c2fe3b117,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:45.407119 kubelet[3195]: E0430 00:38:45.406744 3195 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:45.407119 kubelet[3195]: E0430 00:38:45.406814 3195 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-shvzf" Apr 30 00:38:45.407119 kubelet[3195]: E0430 00:38:45.406833 3195 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-shvzf" Apr 30 00:38:45.407292 kubelet[3195]: E0430 00:38:45.406869 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-shvzf_calico-system(633a0f7e-ae5b-47ef-baa9-358c2fe3b117)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-shvzf_calico-system(633a0f7e-ae5b-47ef-baa9-358c2fe3b117)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:45.408458 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0-shm.mount: Deactivated successfully. Apr 30 00:38:45.443219 kubelet[3195]: I0430 00:38:45.443184 3195 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:38:45.445577 containerd[1704]: time="2025-04-30T00:38:45.444037755Z" level=info msg="StopPodSandbox for \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\"" Apr 30 00:38:45.445577 containerd[1704]: time="2025-04-30T00:38:45.444249675Z" level=info msg="Ensure that sandbox 9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0 in task-service has been cleanup successfully" Apr 30 00:38:45.453516 kubelet[3195]: I0430 00:38:45.453105 3195 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:38:45.454719 containerd[1704]: time="2025-04-30T00:38:45.454373365Z" level=info msg="StopPodSandbox for \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\"" Apr 30 00:38:45.454719 containerd[1704]: time="2025-04-30T00:38:45.454534006Z" level=info msg="Ensure that sandbox c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225 in task-service has been cleanup successfully" Apr 30 00:38:45.456931 kubelet[3195]: I0430 00:38:45.456909 3195 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:38:45.459204 kubelet[3195]: I0430 00:38:45.459181 3195 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:38:45.460087 containerd[1704]: time="2025-04-30T00:38:45.459681131Z" level=info msg="StopPodSandbox for \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\"" Apr 30 00:38:45.460973 containerd[1704]: time="2025-04-30T00:38:45.460941892Z" level=info msg="Ensure that sandbox d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a in task-service has been cleanup successfully" Apr 30 00:38:45.462237 containerd[1704]: time="2025-04-30T00:38:45.461333972Z" level=info msg="StopPodSandbox for \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\"" Apr 30 00:38:45.462935 containerd[1704]: time="2025-04-30T00:38:45.462858974Z" level=info msg="Ensure that sandbox 6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581 in task-service has been cleanup successfully" Apr 30 00:38:45.523256 containerd[1704]: time="2025-04-30T00:38:45.522991755Z" level=error msg="StopPodSandbox for \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\" failed" error="failed to destroy network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:45.523840 kubelet[3195]: E0430 00:38:45.523645 3195 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:38:45.523840 kubelet[3195]: E0430 00:38:45.523719 3195 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225"} Apr 30 00:38:45.523840 kubelet[3195]: E0430 00:38:45.523780 3195 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"2d0e9630-4ffc-4065-9430-75363f7abec1\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 00:38:45.523840 kubelet[3195]: E0430 00:38:45.523811 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"2d0e9630-4ffc-4065-9430-75363f7abec1\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-bgzm6" podUID="2d0e9630-4ffc-4065-9430-75363f7abec1" Apr 30 00:38:45.530161 containerd[1704]: time="2025-04-30T00:38:45.530108002Z" level=error msg="StopPodSandbox for \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\" failed" error="failed to destroy network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:45.530374 kubelet[3195]: E0430 00:38:45.530324 3195 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:38:45.530429 kubelet[3195]: E0430 00:38:45.530372 3195 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a"} Apr 30 00:38:45.530429 kubelet[3195]: E0430 00:38:45.530407 3195 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"8f15ee70-a765-40e4-8bb0-3123a84281b8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 00:38:45.530515 kubelet[3195]: E0430 00:38:45.530432 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"8f15ee70-a765-40e4-8bb0-3123a84281b8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-768b5d7cdb-g4mjx" podUID="8f15ee70-a765-40e4-8bb0-3123a84281b8" Apr 30 00:38:45.531795 containerd[1704]: time="2025-04-30T00:38:45.531679803Z" level=error msg="StopPodSandbox for \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\" failed" error="failed to destroy network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:45.531962 kubelet[3195]: E0430 00:38:45.531902 3195 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:38:45.532006 kubelet[3195]: E0430 00:38:45.531970 3195 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0"} Apr 30 00:38:45.532031 kubelet[3195]: E0430 00:38:45.532010 3195 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"633a0f7e-ae5b-47ef-baa9-358c2fe3b117\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 00:38:45.532114 kubelet[3195]: E0430 00:38:45.532030 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"633a0f7e-ae5b-47ef-baa9-358c2fe3b117\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-shvzf" podUID="633a0f7e-ae5b-47ef-baa9-358c2fe3b117" Apr 30 00:38:45.532611 containerd[1704]: time="2025-04-30T00:38:45.532568924Z" level=error msg="StopPodSandbox for \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\" failed" error="failed to destroy network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:45.532854 kubelet[3195]: E0430 00:38:45.532820 3195 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:38:45.532925 kubelet[3195]: E0430 00:38:45.532856 3195 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581"} Apr 30 00:38:45.532925 kubelet[3195]: E0430 00:38:45.532884 3195 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f93f7ee2-841a-4f1b-a68e-709e05dec366\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 00:38:45.532925 kubelet[3195]: E0430 00:38:45.532911 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f93f7ee2-841a-4f1b-a68e-709e05dec366\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-7z7rd" podUID="f93f7ee2-841a-4f1b-a68e-709e05dec366" Apr 30 00:38:46.122542 containerd[1704]: time="2025-04-30T00:38:46.122490400Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7dd664d894-lcc4f,Uid:1be863bb-f182-4a93-978b-d215d3fc0c5d,Namespace:calico-apiserver,Attempt:0,}" Apr 30 00:38:46.129838 containerd[1704]: time="2025-04-30T00:38:46.129579727Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7dd664d894-tvxc5,Uid:f43c40a3-2edd-4744-8016-2953f9124512,Namespace:calico-apiserver,Attempt:0,}" Apr 30 00:38:46.926669 containerd[1704]: time="2025-04-30T00:38:46.926558291Z" level=error msg="Failed to destroy network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:46.927609 containerd[1704]: time="2025-04-30T00:38:46.927358372Z" level=error msg="encountered an error cleaning up failed sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:46.927609 containerd[1704]: time="2025-04-30T00:38:46.927428972Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7dd664d894-lcc4f,Uid:1be863bb-f182-4a93-978b-d215d3fc0c5d,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:46.929347 kubelet[3195]: E0430 00:38:46.929229 3195 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:46.929347 kubelet[3195]: E0430 00:38:46.929291 3195 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7dd664d894-lcc4f" Apr 30 00:38:46.929347 kubelet[3195]: E0430 00:38:46.929311 3195 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7dd664d894-lcc4f" Apr 30 00:38:46.930526 kubelet[3195]: E0430 00:38:46.929352 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7dd664d894-lcc4f_calico-apiserver(1be863bb-f182-4a93-978b-d215d3fc0c5d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7dd664d894-lcc4f_calico-apiserver(1be863bb-f182-4a93-978b-d215d3fc0c5d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7dd664d894-lcc4f" podUID="1be863bb-f182-4a93-978b-d215d3fc0c5d" Apr 30 00:38:46.933427 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a-shm.mount: Deactivated successfully. Apr 30 00:38:46.941717 containerd[1704]: time="2025-04-30T00:38:46.941631787Z" level=error msg="Failed to destroy network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:46.945351 containerd[1704]: time="2025-04-30T00:38:46.942146627Z" level=error msg="encountered an error cleaning up failed sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:46.945351 containerd[1704]: time="2025-04-30T00:38:46.942240747Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7dd664d894-tvxc5,Uid:f43c40a3-2edd-4744-8016-2953f9124512,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:46.945491 kubelet[3195]: E0430 00:38:46.942531 3195 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:46.945491 kubelet[3195]: E0430 00:38:46.942584 3195 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7dd664d894-tvxc5" Apr 30 00:38:46.945491 kubelet[3195]: E0430 00:38:46.942610 3195 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7dd664d894-tvxc5" Apr 30 00:38:46.945643 kubelet[3195]: E0430 00:38:46.942647 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7dd664d894-tvxc5_calico-apiserver(f43c40a3-2edd-4744-8016-2953f9124512)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7dd664d894-tvxc5_calico-apiserver(f43c40a3-2edd-4744-8016-2953f9124512)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7dd664d894-tvxc5" podUID="f43c40a3-2edd-4744-8016-2953f9124512" Apr 30 00:38:46.948623 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2-shm.mount: Deactivated successfully. Apr 30 00:38:47.465542 kubelet[3195]: I0430 00:38:47.465507 3195 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:38:47.468471 containerd[1704]: time="2025-04-30T00:38:47.468402118Z" level=info msg="StopPodSandbox for \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\"" Apr 30 00:38:47.468668 containerd[1704]: time="2025-04-30T00:38:47.468620358Z" level=info msg="Ensure that sandbox dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2 in task-service has been cleanup successfully" Apr 30 00:38:47.470794 kubelet[3195]: I0430 00:38:47.470354 3195 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:38:47.473525 containerd[1704]: time="2025-04-30T00:38:47.471779482Z" level=info msg="StopPodSandbox for \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\"" Apr 30 00:38:47.473525 containerd[1704]: time="2025-04-30T00:38:47.471987122Z" level=info msg="Ensure that sandbox 2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a in task-service has been cleanup successfully" Apr 30 00:38:47.518700 containerd[1704]: time="2025-04-30T00:38:47.518626569Z" level=error msg="StopPodSandbox for \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\" failed" error="failed to destroy network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:47.519009 kubelet[3195]: E0430 00:38:47.518973 3195 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:38:47.519268 kubelet[3195]: E0430 00:38:47.519153 3195 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a"} Apr 30 00:38:47.519268 kubelet[3195]: E0430 00:38:47.519204 3195 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"1be863bb-f182-4a93-978b-d215d3fc0c5d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 00:38:47.519268 kubelet[3195]: E0430 00:38:47.519234 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"1be863bb-f182-4a93-978b-d215d3fc0c5d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7dd664d894-lcc4f" podUID="1be863bb-f182-4a93-978b-d215d3fc0c5d" Apr 30 00:38:47.524673 containerd[1704]: time="2025-04-30T00:38:47.524607095Z" level=error msg="StopPodSandbox for \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\" failed" error="failed to destroy network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 00:38:47.525104 kubelet[3195]: E0430 00:38:47.524960 3195 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:38:47.525235 kubelet[3195]: E0430 00:38:47.525168 3195 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2"} Apr 30 00:38:47.525283 kubelet[3195]: E0430 00:38:47.525233 3195 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f43c40a3-2edd-4744-8016-2953f9124512\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 00:38:47.525283 kubelet[3195]: E0430 00:38:47.525259 3195 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f43c40a3-2edd-4744-8016-2953f9124512\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7dd664d894-tvxc5" podUID="f43c40a3-2edd-4744-8016-2953f9124512" Apr 30 00:38:48.945818 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2731625257.mount: Deactivated successfully. Apr 30 00:38:50.059022 containerd[1704]: time="2025-04-30T00:38:50.058827493Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:50.063284 containerd[1704]: time="2025-04-30T00:38:50.063204177Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.3: active requests=0, bytes read=138981893" Apr 30 00:38:50.071672 containerd[1704]: time="2025-04-30T00:38:50.071426145Z" level=info msg="ImageCreate event name:\"sha256:cdcce3ec4624a24c28cdc07b0ee29ddf6703628edee7452a3f8a8b4816bfd057\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:50.089331 containerd[1704]: time="2025-04-30T00:38:50.089246523Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:38:50.090504 containerd[1704]: time="2025-04-30T00:38:50.089989204Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.3\" with image id \"sha256:cdcce3ec4624a24c28cdc07b0ee29ddf6703628edee7452a3f8a8b4816bfd057\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\", size \"138981755\" in 5.6471609s" Apr 30 00:38:50.090504 containerd[1704]: time="2025-04-30T00:38:50.090040484Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\" returns image reference \"sha256:cdcce3ec4624a24c28cdc07b0ee29ddf6703628edee7452a3f8a8b4816bfd057\"" Apr 30 00:38:50.104623 containerd[1704]: time="2025-04-30T00:38:50.104577539Z" level=info msg="CreateContainer within sandbox \"90bb9a1599b9e47c9d36951ddbcd3aae8700f36cb9db9f2e7ba366fe64631460\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Apr 30 00:38:50.190758 containerd[1704]: time="2025-04-30T00:38:50.190648026Z" level=info msg="CreateContainer within sandbox \"90bb9a1599b9e47c9d36951ddbcd3aae8700f36cb9db9f2e7ba366fe64631460\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"cdcaefa1cf2354133d788af5005110c406247f63aab7d224d35aff04db810fdd\"" Apr 30 00:38:50.191972 containerd[1704]: time="2025-04-30T00:38:50.191860707Z" level=info msg="StartContainer for \"cdcaefa1cf2354133d788af5005110c406247f63aab7d224d35aff04db810fdd\"" Apr 30 00:38:50.230331 systemd[1]: Started cri-containerd-cdcaefa1cf2354133d788af5005110c406247f63aab7d224d35aff04db810fdd.scope - libcontainer container cdcaefa1cf2354133d788af5005110c406247f63aab7d224d35aff04db810fdd. Apr 30 00:38:50.286049 containerd[1704]: time="2025-04-30T00:38:50.285945962Z" level=info msg="StartContainer for \"cdcaefa1cf2354133d788af5005110c406247f63aab7d224d35aff04db810fdd\" returns successfully" Apr 30 00:38:50.511909 kubelet[3195]: I0430 00:38:50.511830 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-9jgff" podStartSLOduration=2.193249178 podStartE2EDuration="32.51180931s" podCreationTimestamp="2025-04-30 00:38:18 +0000 UTC" firstStartedPulling="2025-04-30 00:38:19.772264593 +0000 UTC m=+14.607490761" lastFinishedPulling="2025-04-30 00:38:50.090824725 +0000 UTC m=+44.926050893" observedRunningTime="2025-04-30 00:38:50.51173663 +0000 UTC m=+45.346962798" watchObservedRunningTime="2025-04-30 00:38:50.51180931 +0000 UTC m=+45.347035478" Apr 30 00:38:50.533789 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Apr 30 00:38:50.533922 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Apr 30 00:38:51.502770 systemd[1]: run-containerd-runc-k8s.io-cdcaefa1cf2354133d788af5005110c406247f63aab7d224d35aff04db810fdd-runc.LiV4je.mount: Deactivated successfully. Apr 30 00:38:52.724194 kernel: bpftool[4526]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Apr 30 00:38:53.037449 systemd-networkd[1329]: vxlan.calico: Link UP Apr 30 00:38:53.037456 systemd-networkd[1329]: vxlan.calico: Gained carrier Apr 30 00:38:54.178259 systemd-networkd[1329]: vxlan.calico: Gained IPv6LL Apr 30 00:38:56.274771 containerd[1704]: time="2025-04-30T00:38:56.274361248Z" level=info msg="StopPodSandbox for \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\"" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.360 [INFO][4612] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.361 [INFO][4612] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" iface="eth0" netns="/var/run/netns/cni-092db076-3486-7f14-8eff-b03a893d5fca" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.361 [INFO][4612] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" iface="eth0" netns="/var/run/netns/cni-092db076-3486-7f14-8eff-b03a893d5fca" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.362 [INFO][4612] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" iface="eth0" netns="/var/run/netns/cni-092db076-3486-7f14-8eff-b03a893d5fca" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.362 [INFO][4612] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.362 [INFO][4612] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.385 [INFO][4619] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" HandleID="k8s-pod-network.9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.386 [INFO][4619] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.386 [INFO][4619] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.395 [WARNING][4619] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" HandleID="k8s-pod-network.9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.395 [INFO][4619] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" HandleID="k8s-pod-network.9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.397 [INFO][4619] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:38:56.402249 containerd[1704]: 2025-04-30 00:38:56.400 [INFO][4612] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:38:56.404256 containerd[1704]: time="2025-04-30T00:38:56.404196002Z" level=info msg="TearDown network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\" successfully" Apr 30 00:38:56.404256 containerd[1704]: time="2025-04-30T00:38:56.404239762Z" level=info msg="StopPodSandbox for \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\" returns successfully" Apr 30 00:38:56.406954 containerd[1704]: time="2025-04-30T00:38:56.406909764Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-shvzf,Uid:633a0f7e-ae5b-47ef-baa9-358c2fe3b117,Namespace:calico-system,Attempt:1,}" Apr 30 00:38:56.407848 systemd[1]: run-netns-cni\x2d092db076\x2d3486\x2d7f14\x2d8eff\x2db03a893d5fca.mount: Deactivated successfully. Apr 30 00:38:56.909195 systemd-networkd[1329]: calicdda35c3166: Link UP Apr 30 00:38:56.911049 systemd-networkd[1329]: calicdda35c3166: Gained carrier Apr 30 00:38:56.925658 systemd[1]: run-containerd-runc-k8s.io-cdcaefa1cf2354133d788af5005110c406247f63aab7d224d35aff04db810fdd-runc.OmpXbA.mount: Deactivated successfully. Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.711 [INFO][4630] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0 csi-node-driver- calico-system 633a0f7e-ae5b-47ef-baa9-358c2fe3b117 782 0 2025-04-30 00:38:19 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:5bcd8f69 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4081.3.3-a-c388cdac8a csi-node-driver-shvzf eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calicdda35c3166 [] []}} ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Namespace="calico-system" Pod="csi-node-driver-shvzf" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.711 [INFO][4630] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Namespace="calico-system" Pod="csi-node-driver-shvzf" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.743 [INFO][4637] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" HandleID="k8s-pod-network.487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.761 [INFO][4637] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" HandleID="k8s-pod-network.487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000333520), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.3-a-c388cdac8a", "pod":"csi-node-driver-shvzf", "timestamp":"2025-04-30 00:38:56.743539979 +0000 UTC"}, Hostname:"ci-4081.3.3-a-c388cdac8a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.761 [INFO][4637] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.761 [INFO][4637] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.761 [INFO][4637] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.3-a-c388cdac8a' Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.764 [INFO][4637] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.769 [INFO][4637] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.774 [INFO][4637] ipam/ipam.go 489: Trying affinity for 192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.777 [INFO][4637] ipam/ipam.go 155: Attempting to load block cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.780 [INFO][4637] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.780 [INFO][4637] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.17.128/26 handle="k8s-pod-network.487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.782 [INFO][4637] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.892 [INFO][4637] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.17.128/26 handle="k8s-pod-network.487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.902 [INFO][4637] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.17.129/26] block=192.168.17.128/26 handle="k8s-pod-network.487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.902 [INFO][4637] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.17.129/26] handle="k8s-pod-network.487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.902 [INFO][4637] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:38:56.956143 containerd[1704]: 2025-04-30 00:38:56.902 [INFO][4637] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.17.129/26] IPv6=[] ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" HandleID="k8s-pod-network.487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:56.957013 containerd[1704]: 2025-04-30 00:38:56.905 [INFO][4630] cni-plugin/k8s.go 386: Populated endpoint ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Namespace="calico-system" Pod="csi-node-driver-shvzf" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"633a0f7e-ae5b-47ef-baa9-358c2fe3b117", ResourceVersion:"782", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5bcd8f69", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"", Pod:"csi-node-driver-shvzf", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.17.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calicdda35c3166", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:38:56.957013 containerd[1704]: 2025-04-30 00:38:56.905 [INFO][4630] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.17.129/32] ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Namespace="calico-system" Pod="csi-node-driver-shvzf" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:56.957013 containerd[1704]: 2025-04-30 00:38:56.905 [INFO][4630] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicdda35c3166 ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Namespace="calico-system" Pod="csi-node-driver-shvzf" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:56.957013 containerd[1704]: 2025-04-30 00:38:56.912 [INFO][4630] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Namespace="calico-system" Pod="csi-node-driver-shvzf" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:56.957013 containerd[1704]: 2025-04-30 00:38:56.913 [INFO][4630] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Namespace="calico-system" Pod="csi-node-driver-shvzf" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"633a0f7e-ae5b-47ef-baa9-358c2fe3b117", ResourceVersion:"782", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5bcd8f69", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c", Pod:"csi-node-driver-shvzf", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.17.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calicdda35c3166", MAC:"46:8a:8d:d5:de:2b", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:38:56.957013 containerd[1704]: 2025-04-30 00:38:56.945 [INFO][4630] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c" Namespace="calico-system" Pod="csi-node-driver-shvzf" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:38:57.025516 containerd[1704]: time="2025-04-30T00:38:57.025303745Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:38:57.025516 containerd[1704]: time="2025-04-30T00:38:57.025376785Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:38:57.025516 containerd[1704]: time="2025-04-30T00:38:57.025391585Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:57.025895 containerd[1704]: time="2025-04-30T00:38:57.025495225Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:57.045806 systemd[1]: Started cri-containerd-487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c.scope - libcontainer container 487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c. Apr 30 00:38:57.085681 containerd[1704]: time="2025-04-30T00:38:57.085632678Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-shvzf,Uid:633a0f7e-ae5b-47ef-baa9-358c2fe3b117,Namespace:calico-system,Attempt:1,} returns sandbox id \"487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c\"" Apr 30 00:38:57.088485 containerd[1704]: time="2025-04-30T00:38:57.088354520Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\"" Apr 30 00:38:57.275957 containerd[1704]: time="2025-04-30T00:38:57.275822604Z" level=info msg="StopPodSandbox for \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\"" Apr 30 00:38:57.277167 containerd[1704]: time="2025-04-30T00:38:57.277018045Z" level=info msg="StopPodSandbox for \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\"" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.383 [INFO][4745] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.383 [INFO][4745] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" iface="eth0" netns="/var/run/netns/cni-710a744f-9bf8-8866-a9c7-f85b871ccaec" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.384 [INFO][4745] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" iface="eth0" netns="/var/run/netns/cni-710a744f-9bf8-8866-a9c7-f85b871ccaec" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.384 [INFO][4745] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" iface="eth0" netns="/var/run/netns/cni-710a744f-9bf8-8866-a9c7-f85b871ccaec" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.384 [INFO][4745] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.384 [INFO][4745] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.412 [INFO][4764] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" HandleID="k8s-pod-network.6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.412 [INFO][4764] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.412 [INFO][4764] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.425 [WARNING][4764] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" HandleID="k8s-pod-network.6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.427 [INFO][4764] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" HandleID="k8s-pod-network.6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.430 [INFO][4764] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:38:57.434591 containerd[1704]: 2025-04-30 00:38:57.432 [INFO][4745] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:38:57.435290 containerd[1704]: time="2025-04-30T00:38:57.435235224Z" level=info msg="TearDown network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\" successfully" Apr 30 00:38:57.435290 containerd[1704]: time="2025-04-30T00:38:57.435285984Z" level=info msg="StopPodSandbox for \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\" returns successfully" Apr 30 00:38:57.440697 containerd[1704]: time="2025-04-30T00:38:57.440418548Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-7z7rd,Uid:f93f7ee2-841a-4f1b-a68e-709e05dec366,Namespace:kube-system,Attempt:1,}" Apr 30 00:38:57.441571 systemd[1]: run-netns-cni\x2d710a744f\x2d9bf8\x2d8866\x2da9c7\x2df85b871ccaec.mount: Deactivated successfully. Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.426 [INFO][4753] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.427 [INFO][4753] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" iface="eth0" netns="/var/run/netns/cni-f729fc2e-8543-d806-a449-20afaf17478f" Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.427 [INFO][4753] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" iface="eth0" netns="/var/run/netns/cni-f729fc2e-8543-d806-a449-20afaf17478f" Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.428 [INFO][4753] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" iface="eth0" netns="/var/run/netns/cni-f729fc2e-8543-d806-a449-20afaf17478f" Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.428 [INFO][4753] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.428 [INFO][4753] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.462 [INFO][4771] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" HandleID="k8s-pod-network.c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.463 [INFO][4771] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.463 [INFO][4771] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.474 [WARNING][4771] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" HandleID="k8s-pod-network.c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.474 [INFO][4771] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" HandleID="k8s-pod-network.c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.476 [INFO][4771] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:38:57.480719 containerd[1704]: 2025-04-30 00:38:57.478 [INFO][4753] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:38:57.481503 containerd[1704]: time="2025-04-30T00:38:57.481391344Z" level=info msg="TearDown network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\" successfully" Apr 30 00:38:57.481503 containerd[1704]: time="2025-04-30T00:38:57.481425904Z" level=info msg="StopPodSandbox for \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\" returns successfully" Apr 30 00:38:57.484093 containerd[1704]: time="2025-04-30T00:38:57.482350385Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-bgzm6,Uid:2d0e9630-4ffc-4065-9430-75363f7abec1,Namespace:kube-system,Attempt:1,}" Apr 30 00:38:57.484862 systemd[1]: run-netns-cni\x2df729fc2e\x2d8543\x2dd806\x2da449\x2d20afaf17478f.mount: Deactivated successfully. Apr 30 00:38:58.389642 systemd-networkd[1329]: cali7b9e637ad69: Link UP Apr 30 00:38:58.389812 systemd-networkd[1329]: cali7b9e637ad69: Gained carrier Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.299 [INFO][4777] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0 coredns-6f6b679f8f- kube-system f93f7ee2-841a-4f1b-a68e-709e05dec366 794 0 2025-04-30 00:38:09 +0000 UTC map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.3-a-c388cdac8a coredns-6f6b679f8f-7z7rd eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali7b9e637ad69 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Namespace="kube-system" Pod="coredns-6f6b679f8f-7z7rd" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.300 [INFO][4777] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Namespace="kube-system" Pod="coredns-6f6b679f8f-7z7rd" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.331 [INFO][4790] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" HandleID="k8s-pod-network.dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.344 [INFO][4790] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" HandleID="k8s-pod-network.dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003318e0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.3-a-c388cdac8a", "pod":"coredns-6f6b679f8f-7z7rd", "timestamp":"2025-04-30 00:38:58.331408128 +0000 UTC"}, Hostname:"ci-4081.3.3-a-c388cdac8a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.344 [INFO][4790] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.344 [INFO][4790] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.344 [INFO][4790] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.3-a-c388cdac8a' Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.347 [INFO][4790] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.351 [INFO][4790] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.357 [INFO][4790] ipam/ipam.go 489: Trying affinity for 192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.359 [INFO][4790] ipam/ipam.go 155: Attempting to load block cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.362 [INFO][4790] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.362 [INFO][4790] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.17.128/26 handle="k8s-pod-network.dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.364 [INFO][4790] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.375 [INFO][4790] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.17.128/26 handle="k8s-pod-network.dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.383 [INFO][4790] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.17.130/26] block=192.168.17.128/26 handle="k8s-pod-network.dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.384 [INFO][4790] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.17.130/26] handle="k8s-pod-network.dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.384 [INFO][4790] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:38:58.442274 containerd[1704]: 2025-04-30 00:38:58.384 [INFO][4790] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.17.130/26] IPv6=[] ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" HandleID="k8s-pod-network.dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:58.446180 containerd[1704]: 2025-04-30 00:38:58.386 [INFO][4777] cni-plugin/k8s.go 386: Populated endpoint ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Namespace="kube-system" Pod="coredns-6f6b679f8f-7z7rd" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"f93f7ee2-841a-4f1b-a68e-709e05dec366", ResourceVersion:"794", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"", Pod:"coredns-6f6b679f8f-7z7rd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.17.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7b9e637ad69", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:38:58.446180 containerd[1704]: 2025-04-30 00:38:58.386 [INFO][4777] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.17.130/32] ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Namespace="kube-system" Pod="coredns-6f6b679f8f-7z7rd" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:58.446180 containerd[1704]: 2025-04-30 00:38:58.386 [INFO][4777] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7b9e637ad69 ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Namespace="kube-system" Pod="coredns-6f6b679f8f-7z7rd" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:58.446180 containerd[1704]: 2025-04-30 00:38:58.390 [INFO][4777] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Namespace="kube-system" Pod="coredns-6f6b679f8f-7z7rd" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:58.446180 containerd[1704]: 2025-04-30 00:38:58.391 [INFO][4777] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Namespace="kube-system" Pod="coredns-6f6b679f8f-7z7rd" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"f93f7ee2-841a-4f1b-a68e-709e05dec366", ResourceVersion:"794", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec", Pod:"coredns-6f6b679f8f-7z7rd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.17.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7b9e637ad69", MAC:"76:e9:df:e0:45:e9", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:38:58.446180 containerd[1704]: 2025-04-30 00:38:58.438 [INFO][4777] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec" Namespace="kube-system" Pod="coredns-6f6b679f8f-7z7rd" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:38:58.536158 containerd[1704]: time="2025-04-30T00:38:58.535666346Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:38:58.536158 containerd[1704]: time="2025-04-30T00:38:58.535747027Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:38:58.536158 containerd[1704]: time="2025-04-30T00:38:58.535762347Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:58.536158 containerd[1704]: time="2025-04-30T00:38:58.535853467Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:58.573333 systemd[1]: Started cri-containerd-dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec.scope - libcontainer container dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec. Apr 30 00:38:58.627739 containerd[1704]: time="2025-04-30T00:38:58.627643027Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-7z7rd,Uid:f93f7ee2-841a-4f1b-a68e-709e05dec366,Namespace:kube-system,Attempt:1,} returns sandbox id \"dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec\"" Apr 30 00:38:58.636439 containerd[1704]: time="2025-04-30T00:38:58.636370955Z" level=info msg="CreateContainer within sandbox \"dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 30 00:38:58.694277 systemd-networkd[1329]: cali989970102d8: Link UP Apr 30 00:38:58.694460 systemd-networkd[1329]: cali989970102d8: Gained carrier Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.519 [INFO][4813] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0 coredns-6f6b679f8f- kube-system 2d0e9630-4ffc-4065-9430-75363f7abec1 795 0 2025-04-30 00:38:09 +0000 UTC map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.3-a-c388cdac8a coredns-6f6b679f8f-bgzm6 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali989970102d8 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bgzm6" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.519 [INFO][4813] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bgzm6" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.574 [INFO][4840] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" HandleID="k8s-pod-network.3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.593 [INFO][4840] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" HandleID="k8s-pod-network.3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000384a40), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.3-a-c388cdac8a", "pod":"coredns-6f6b679f8f-bgzm6", "timestamp":"2025-04-30 00:38:58.57415914 +0000 UTC"}, Hostname:"ci-4081.3.3-a-c388cdac8a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.593 [INFO][4840] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.593 [INFO][4840] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.593 [INFO][4840] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.3-a-c388cdac8a' Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.596 [INFO][4840] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.604 [INFO][4840] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.614 [INFO][4840] ipam/ipam.go 489: Trying affinity for 192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.619 [INFO][4840] ipam/ipam.go 155: Attempting to load block cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.624 [INFO][4840] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.624 [INFO][4840] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.17.128/26 handle="k8s-pod-network.3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.630 [INFO][4840] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.640 [INFO][4840] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.17.128/26 handle="k8s-pod-network.3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.686 [INFO][4840] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.17.131/26] block=192.168.17.128/26 handle="k8s-pod-network.3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.687 [INFO][4840] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.17.131/26] handle="k8s-pod-network.3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.687 [INFO][4840] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:38:58.765480 containerd[1704]: 2025-04-30 00:38:58.687 [INFO][4840] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.17.131/26] IPv6=[] ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" HandleID="k8s-pod-network.3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:58.766043 containerd[1704]: 2025-04-30 00:38:58.691 [INFO][4813] cni-plugin/k8s.go 386: Populated endpoint ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bgzm6" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"2d0e9630-4ffc-4065-9430-75363f7abec1", ResourceVersion:"795", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"", Pod:"coredns-6f6b679f8f-bgzm6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.17.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali989970102d8", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:38:58.766043 containerd[1704]: 2025-04-30 00:38:58.691 [INFO][4813] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.17.131/32] ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bgzm6" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:58.766043 containerd[1704]: 2025-04-30 00:38:58.691 [INFO][4813] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali989970102d8 ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bgzm6" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:58.766043 containerd[1704]: 2025-04-30 00:38:58.693 [INFO][4813] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bgzm6" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:58.766043 containerd[1704]: 2025-04-30 00:38:58.693 [INFO][4813] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bgzm6" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"2d0e9630-4ffc-4065-9430-75363f7abec1", ResourceVersion:"795", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a", Pod:"coredns-6f6b679f8f-bgzm6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.17.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali989970102d8", MAC:"b2:9b:64:14:e5:b3", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:38:58.766043 containerd[1704]: 2025-04-30 00:38:58.732 [INFO][4813] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bgzm6" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:38:58.850227 systemd-networkd[1329]: calicdda35c3166: Gained IPv6LL Apr 30 00:38:58.982785 containerd[1704]: time="2025-04-30T00:38:58.982357297Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:38:58.982785 containerd[1704]: time="2025-04-30T00:38:58.982429097Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:38:58.982785 containerd[1704]: time="2025-04-30T00:38:58.982440617Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:58.982785 containerd[1704]: time="2025-04-30T00:38:58.982627457Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:38:59.010367 systemd[1]: Started cri-containerd-3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a.scope - libcontainer container 3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a. Apr 30 00:38:59.049545 containerd[1704]: time="2025-04-30T00:38:59.049489196Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-bgzm6,Uid:2d0e9630-4ffc-4065-9430-75363f7abec1,Namespace:kube-system,Attempt:1,} returns sandbox id \"3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a\"" Apr 30 00:38:59.054550 containerd[1704]: time="2025-04-30T00:38:59.054495240Z" level=info msg="CreateContainer within sandbox \"3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 30 00:38:59.568142 containerd[1704]: time="2025-04-30T00:38:59.567996891Z" level=info msg="CreateContainer within sandbox \"dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"fdfa9f1efafd6b447294b757d771dcc6a0acdc3d4a071bf96f11dee4f8e9f4c6\"" Apr 30 00:38:59.569330 containerd[1704]: time="2025-04-30T00:38:59.569191172Z" level=info msg="StartContainer for \"fdfa9f1efafd6b447294b757d771dcc6a0acdc3d4a071bf96f11dee4f8e9f4c6\"" Apr 30 00:38:59.604300 systemd[1]: Started cri-containerd-fdfa9f1efafd6b447294b757d771dcc6a0acdc3d4a071bf96f11dee4f8e9f4c6.scope - libcontainer container fdfa9f1efafd6b447294b757d771dcc6a0acdc3d4a071bf96f11dee4f8e9f4c6. Apr 30 00:38:59.682269 systemd-networkd[1329]: cali7b9e637ad69: Gained IPv6LL Apr 30 00:38:59.712314 containerd[1704]: time="2025-04-30T00:38:59.712253827Z" level=info msg="StartContainer for \"fdfa9f1efafd6b447294b757d771dcc6a0acdc3d4a071bf96f11dee4f8e9f4c6\" returns successfully" Apr 30 00:39:00.169975 containerd[1704]: time="2025-04-30T00:39:00.169822658Z" level=info msg="CreateContainer within sandbox \"3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"6ad3e69601945d2b65a644440e7619f68fd1e0f79de08c9ed6e615dbdddfb6b7\"" Apr 30 00:39:00.172145 containerd[1704]: time="2025-04-30T00:39:00.171596179Z" level=info msg="StartContainer for \"6ad3e69601945d2b65a644440e7619f68fd1e0f79de08c9ed6e615dbdddfb6b7\"" Apr 30 00:39:00.208383 systemd[1]: Started cri-containerd-6ad3e69601945d2b65a644440e7619f68fd1e0f79de08c9ed6e615dbdddfb6b7.scope - libcontainer container 6ad3e69601945d2b65a644440e7619f68fd1e0f79de08c9ed6e615dbdddfb6b7. Apr 30 00:39:00.275515 containerd[1704]: time="2025-04-30T00:39:00.275461517Z" level=info msg="StopPodSandbox for \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\"" Apr 30 00:39:00.718620 containerd[1704]: time="2025-04-30T00:39:00.275804517Z" level=info msg="StopPodSandbox for \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\"" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.396 [INFO][5031] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.396 [INFO][5031] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" iface="eth0" netns="/var/run/netns/cni-815efc7a-41b4-3c2c-5dee-4e545a13bce4" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.396 [INFO][5031] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" iface="eth0" netns="/var/run/netns/cni-815efc7a-41b4-3c2c-5dee-4e545a13bce4" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.397 [INFO][5031] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" iface="eth0" netns="/var/run/netns/cni-815efc7a-41b4-3c2c-5dee-4e545a13bce4" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.397 [INFO][5031] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.397 [INFO][5031] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.444 [INFO][5046] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" HandleID="k8s-pod-network.d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.444 [INFO][5046] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.444 [INFO][5046] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.457 [WARNING][5046] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" HandleID="k8s-pod-network.d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.457 [INFO][5046] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" HandleID="k8s-pod-network.d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.461 [INFO][5046] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.465 [INFO][5031] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:00.718620 containerd[1704]: time="2025-04-30T00:39:00.468382019Z" level=info msg="TearDown network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\" successfully" Apr 30 00:39:00.718620 containerd[1704]: time="2025-04-30T00:39:00.468410739Z" level=info msg="StopPodSandbox for \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\" returns successfully" Apr 30 00:39:00.718620 containerd[1704]: time="2025-04-30T00:39:00.472833143Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-768b5d7cdb-g4mjx,Uid:8f15ee70-a765-40e4-8bb0-3123a84281b8,Namespace:calico-system,Attempt:1,}" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.402 [INFO][5033] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.403 [INFO][5033] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" iface="eth0" netns="/var/run/netns/cni-3e7ba3db-bf0b-7dfc-d3c9-a198e1db90f3" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.403 [INFO][5033] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" iface="eth0" netns="/var/run/netns/cni-3e7ba3db-bf0b-7dfc-d3c9-a198e1db90f3" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.406 [INFO][5033] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" iface="eth0" netns="/var/run/netns/cni-3e7ba3db-bf0b-7dfc-d3c9-a198e1db90f3" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.406 [INFO][5033] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.406 [INFO][5033] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.456 [INFO][5051] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" HandleID="k8s-pod-network.2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.457 [INFO][5051] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.462 [INFO][5051] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.481 [WARNING][5051] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" HandleID="k8s-pod-network.2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:00.718620 containerd[1704]: 2025-04-30 00:39:00.481 [INFO][5051] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" HandleID="k8s-pod-network.2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:00.475052 systemd[1]: run-netns-cni\x2d815efc7a\x2d41b4\x2d3c2c\x2d5dee\x2d4e545a13bce4.mount: Deactivated successfully. Apr 30 00:39:00.719644 containerd[1704]: 2025-04-30 00:39:00.483 [INFO][5051] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:00.719644 containerd[1704]: 2025-04-30 00:39:00.485 [INFO][5033] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:00.719644 containerd[1704]: time="2025-04-30T00:39:00.490531040Z" level=info msg="TearDown network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\" successfully" Apr 30 00:39:00.719644 containerd[1704]: time="2025-04-30T00:39:00.490570080Z" level=info msg="StopPodSandbox for \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\" returns successfully" Apr 30 00:39:00.719644 containerd[1704]: time="2025-04-30T00:39:00.492858522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7dd664d894-lcc4f,Uid:1be863bb-f182-4a93-978b-d215d3fc0c5d,Namespace:calico-apiserver,Attempt:1,}" Apr 30 00:39:00.492048 systemd[1]: run-netns-cni\x2d3e7ba3db\x2dbf0b\x2d7dfc\x2dd3c9\x2da198e1db90f3.mount: Deactivated successfully. Apr 30 00:39:00.579238 systemd-networkd[1329]: cali989970102d8: Gained IPv6LL Apr 30 00:39:01.267927 containerd[1704]: time="2025-04-30T00:39:01.267826292Z" level=info msg="StartContainer for \"6ad3e69601945d2b65a644440e7619f68fd1e0f79de08c9ed6e615dbdddfb6b7\" returns successfully" Apr 30 00:39:01.321616 kubelet[3195]: I0430 00:39:01.321512 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-7z7rd" podStartSLOduration=52.321489542 podStartE2EDuration="52.321489542s" podCreationTimestamp="2025-04-30 00:38:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 00:39:01.320534861 +0000 UTC m=+56.155761069" watchObservedRunningTime="2025-04-30 00:39:01.321489542 +0000 UTC m=+56.156715710" Apr 30 00:39:02.275137 containerd[1704]: time="2025-04-30T00:39:02.275092001Z" level=info msg="StopPodSandbox for \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\"" Apr 30 00:39:02.404462 kubelet[3195]: I0430 00:39:02.403364 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-bgzm6" podStartSLOduration=53.403343801 podStartE2EDuration="53.403343801s" podCreationTimestamp="2025-04-30 00:38:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 00:39:02.325528448 +0000 UTC m=+57.160754656" watchObservedRunningTime="2025-04-30 00:39:02.403343801 +0000 UTC m=+57.238569969" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.402 [INFO][5083] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.403 [INFO][5083] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" iface="eth0" netns="/var/run/netns/cni-9cd6ccfd-013c-b0e2-9419-87d4b6de5b8d" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.403 [INFO][5083] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" iface="eth0" netns="/var/run/netns/cni-9cd6ccfd-013c-b0e2-9419-87d4b6de5b8d" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.405 [INFO][5083] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" iface="eth0" netns="/var/run/netns/cni-9cd6ccfd-013c-b0e2-9419-87d4b6de5b8d" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.405 [INFO][5083] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.405 [INFO][5083] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.438 [INFO][5092] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" HandleID="k8s-pod-network.dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.438 [INFO][5092] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.438 [INFO][5092] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.454 [WARNING][5092] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" HandleID="k8s-pod-network.dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.454 [INFO][5092] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" HandleID="k8s-pod-network.dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.456 [INFO][5092] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:02.461302 containerd[1704]: 2025-04-30 00:39:02.459 [INFO][5083] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:02.463958 systemd[1]: run-netns-cni\x2d9cd6ccfd\x2d013c\x2db0e2\x2d9419\x2d87d4b6de5b8d.mount: Deactivated successfully. Apr 30 00:39:02.464806 containerd[1704]: time="2025-04-30T00:39:02.464038739Z" level=info msg="TearDown network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\" successfully" Apr 30 00:39:02.464806 containerd[1704]: time="2025-04-30T00:39:02.464150819Z" level=info msg="StopPodSandbox for \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\" returns successfully" Apr 30 00:39:02.466723 containerd[1704]: time="2025-04-30T00:39:02.466411421Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7dd664d894-tvxc5,Uid:f43c40a3-2edd-4744-8016-2953f9124512,Namespace:calico-apiserver,Attempt:1,}" Apr 30 00:39:03.388295 systemd-networkd[1329]: calide548167888: Link UP Apr 30 00:39:03.388592 systemd-networkd[1329]: calide548167888: Gained carrier Apr 30 00:39:03.412326 containerd[1704]: time="2025-04-30T00:39:03.412262472Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.209 [INFO][5098] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0 calico-kube-controllers-768b5d7cdb- calico-system 8f15ee70-a765-40e4-8bb0-3123a84281b8 814 0 2025-04-30 00:38:19 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:768b5d7cdb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4081.3.3-a-c388cdac8a calico-kube-controllers-768b5d7cdb-g4mjx eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calide548167888 [] []}} ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Namespace="calico-system" Pod="calico-kube-controllers-768b5d7cdb-g4mjx" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.209 [INFO][5098] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Namespace="calico-system" Pod="calico-kube-controllers-768b5d7cdb-g4mjx" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.262 [INFO][5111] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" HandleID="k8s-pod-network.cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.279 [INFO][5111] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" HandleID="k8s-pod-network.cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004c140), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.3-a-c388cdac8a", "pod":"calico-kube-controllers-768b5d7cdb-g4mjx", "timestamp":"2025-04-30 00:39:03.262691651 +0000 UTC"}, Hostname:"ci-4081.3.3-a-c388cdac8a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.280 [INFO][5111] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.280 [INFO][5111] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.280 [INFO][5111] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.3-a-c388cdac8a' Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.283 [INFO][5111] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.288 [INFO][5111] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.297 [INFO][5111] ipam/ipam.go 489: Trying affinity for 192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.300 [INFO][5111] ipam/ipam.go 155: Attempting to load block cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.305 [INFO][5111] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.305 [INFO][5111] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.17.128/26 handle="k8s-pod-network.cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.309 [INFO][5111] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0 Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.319 [INFO][5111] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.17.128/26 handle="k8s-pod-network.cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.380 [INFO][5111] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.17.132/26] block=192.168.17.128/26 handle="k8s-pod-network.cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.380 [INFO][5111] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.17.132/26] handle="k8s-pod-network.cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.380 [INFO][5111] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:03.460516 containerd[1704]: 2025-04-30 00:39:03.380 [INFO][5111] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.17.132/26] IPv6=[] ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" HandleID="k8s-pod-network.cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:03.461161 containerd[1704]: 2025-04-30 00:39:03.383 [INFO][5098] cni-plugin/k8s.go 386: Populated endpoint ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Namespace="calico-system" Pod="calico-kube-controllers-768b5d7cdb-g4mjx" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0", GenerateName:"calico-kube-controllers-768b5d7cdb-", Namespace:"calico-system", SelfLink:"", UID:"8f15ee70-a765-40e4-8bb0-3123a84281b8", ResourceVersion:"814", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"768b5d7cdb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"", Pod:"calico-kube-controllers-768b5d7cdb-g4mjx", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.17.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calide548167888", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:03.461161 containerd[1704]: 2025-04-30 00:39:03.383 [INFO][5098] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.17.132/32] ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Namespace="calico-system" Pod="calico-kube-controllers-768b5d7cdb-g4mjx" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:03.461161 containerd[1704]: 2025-04-30 00:39:03.383 [INFO][5098] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calide548167888 ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Namespace="calico-system" Pod="calico-kube-controllers-768b5d7cdb-g4mjx" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:03.461161 containerd[1704]: 2025-04-30 00:39:03.387 [INFO][5098] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Namespace="calico-system" Pod="calico-kube-controllers-768b5d7cdb-g4mjx" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:03.461161 containerd[1704]: 2025-04-30 00:39:03.388 [INFO][5098] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Namespace="calico-system" Pod="calico-kube-controllers-768b5d7cdb-g4mjx" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0", GenerateName:"calico-kube-controllers-768b5d7cdb-", Namespace:"calico-system", SelfLink:"", UID:"8f15ee70-a765-40e4-8bb0-3123a84281b8", ResourceVersion:"814", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"768b5d7cdb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0", Pod:"calico-kube-controllers-768b5d7cdb-g4mjx", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.17.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calide548167888", MAC:"0e:62:99:b4:f5:13", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:03.461161 containerd[1704]: 2025-04-30 00:39:03.416 [INFO][5098] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0" Namespace="calico-system" Pod="calico-kube-controllers-768b5d7cdb-g4mjx" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:03.535023 containerd[1704]: time="2025-04-30T00:39:03.534968427Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.3: active requests=0, bytes read=7474935" Apr 30 00:39:03.648014 containerd[1704]: time="2025-04-30T00:39:03.647263213Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:39:03.648014 containerd[1704]: time="2025-04-30T00:39:03.647807534Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:39:03.648014 containerd[1704]: time="2025-04-30T00:39:03.647825614Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:39:03.648894 containerd[1704]: time="2025-04-30T00:39:03.648561334Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:39:03.665781 containerd[1704]: time="2025-04-30T00:39:03.665691430Z" level=info msg="ImageCreate event name:\"sha256:15faf29e8b518d846c91c15785ff89e783d356ea0f2b22826f47a556ea32645b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:03.691343 systemd[1]: Started cri-containerd-cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0.scope - libcontainer container cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0. Apr 30 00:39:03.712924 systemd-networkd[1329]: calieecd7fb4aeb: Link UP Apr 30 00:39:03.717442 systemd-networkd[1329]: calieecd7fb4aeb: Gained carrier Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.549 [INFO][5128] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0 calico-apiserver-7dd664d894- calico-apiserver 1be863bb-f182-4a93-978b-d215d3fc0c5d 815 0 2025-04-30 00:38:18 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7dd664d894 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.3-a-c388cdac8a calico-apiserver-7dd664d894-lcc4f eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calieecd7fb4aeb [] []}} ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-lcc4f" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.549 [INFO][5128] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-lcc4f" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.583 [INFO][5153] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" HandleID="k8s-pod-network.d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.599 [INFO][5153] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" HandleID="k8s-pod-network.d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000316b00), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.3-a-c388cdac8a", "pod":"calico-apiserver-7dd664d894-lcc4f", "timestamp":"2025-04-30 00:39:03.583037433 +0000 UTC"}, Hostname:"ci-4081.3.3-a-c388cdac8a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.599 [INFO][5153] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.599 [INFO][5153] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.599 [INFO][5153] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.3-a-c388cdac8a' Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.601 [INFO][5153] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.610 [INFO][5153] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.632 [INFO][5153] ipam/ipam.go 489: Trying affinity for 192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.635 [INFO][5153] ipam/ipam.go 155: Attempting to load block cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.644 [INFO][5153] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.646 [INFO][5153] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.17.128/26 handle="k8s-pod-network.d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.650 [INFO][5153] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858 Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.665 [INFO][5153] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.17.128/26 handle="k8s-pod-network.d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.686 [INFO][5153] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.17.133/26] block=192.168.17.128/26 handle="k8s-pod-network.d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.687 [INFO][5153] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.17.133/26] handle="k8s-pod-network.d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.687 [INFO][5153] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:03.769315 containerd[1704]: 2025-04-30 00:39:03.687 [INFO][5153] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.17.133/26] IPv6=[] ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" HandleID="k8s-pod-network.d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:03.769990 containerd[1704]: 2025-04-30 00:39:03.699 [INFO][5128] cni-plugin/k8s.go 386: Populated endpoint ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-lcc4f" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0", GenerateName:"calico-apiserver-7dd664d894-", Namespace:"calico-apiserver", SelfLink:"", UID:"1be863bb-f182-4a93-978b-d215d3fc0c5d", ResourceVersion:"815", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7dd664d894", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"", Pod:"calico-apiserver-7dd664d894-lcc4f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.17.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieecd7fb4aeb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:03.769990 containerd[1704]: 2025-04-30 00:39:03.701 [INFO][5128] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.17.133/32] ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-lcc4f" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:03.769990 containerd[1704]: 2025-04-30 00:39:03.701 [INFO][5128] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calieecd7fb4aeb ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-lcc4f" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:03.769990 containerd[1704]: 2025-04-30 00:39:03.716 [INFO][5128] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-lcc4f" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:03.769990 containerd[1704]: 2025-04-30 00:39:03.716 [INFO][5128] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-lcc4f" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0", GenerateName:"calico-apiserver-7dd664d894-", Namespace:"calico-apiserver", SelfLink:"", UID:"1be863bb-f182-4a93-978b-d215d3fc0c5d", ResourceVersion:"815", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7dd664d894", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858", Pod:"calico-apiserver-7dd664d894-lcc4f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.17.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieecd7fb4aeb", MAC:"0a:ad:24:25:15:de", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:03.769990 containerd[1704]: 2025-04-30 00:39:03.754 [INFO][5128] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-lcc4f" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:03.821376 containerd[1704]: time="2025-04-30T00:39:03.820831496Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:03.823389 containerd[1704]: time="2025-04-30T00:39:03.822916018Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.3\" with image id \"sha256:15faf29e8b518d846c91c15785ff89e783d356ea0f2b22826f47a556ea32645b\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\", size \"8844117\" in 6.734487978s" Apr 30 00:39:03.823389 containerd[1704]: time="2025-04-30T00:39:03.822981779Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\" returns image reference \"sha256:15faf29e8b518d846c91c15785ff89e783d356ea0f2b22826f47a556ea32645b\"" Apr 30 00:39:03.849449 containerd[1704]: time="2025-04-30T00:39:03.849407283Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-768b5d7cdb-g4mjx,Uid:8f15ee70-a765-40e4-8bb0-3123a84281b8,Namespace:calico-system,Attempt:1,} returns sandbox id \"cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0\"" Apr 30 00:39:03.862871 containerd[1704]: time="2025-04-30T00:39:03.862622096Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\"" Apr 30 00:39:03.862871 containerd[1704]: time="2025-04-30T00:39:03.862631616Z" level=info msg="CreateContainer within sandbox \"487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Apr 30 00:39:03.906496 containerd[1704]: time="2025-04-30T00:39:03.906014857Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:39:03.906496 containerd[1704]: time="2025-04-30T00:39:03.906124017Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:39:03.906496 containerd[1704]: time="2025-04-30T00:39:03.906135497Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:39:03.907205 containerd[1704]: time="2025-04-30T00:39:03.906975698Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:39:03.935327 systemd[1]: Started cri-containerd-d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858.scope - libcontainer container d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858. Apr 30 00:39:03.996535 containerd[1704]: time="2025-04-30T00:39:03.996480022Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7dd664d894-lcc4f,Uid:1be863bb-f182-4a93-978b-d215d3fc0c5d,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858\"" Apr 30 00:39:04.027774 systemd-networkd[1329]: calia26b2b102cd: Link UP Apr 30 00:39:04.028584 systemd-networkd[1329]: calia26b2b102cd: Gained carrier Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.802 [INFO][5197] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0 calico-apiserver-7dd664d894- calico-apiserver f43c40a3-2edd-4744-8016-2953f9124512 836 0 2025-04-30 00:38:18 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7dd664d894 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.3-a-c388cdac8a calico-apiserver-7dd664d894-tvxc5 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calia26b2b102cd [] []}} ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-tvxc5" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.802 [INFO][5197] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-tvxc5" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.898 [INFO][5219] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" HandleID="k8s-pod-network.42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.942 [INFO][5219] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" HandleID="k8s-pod-network.42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000123880), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.3-a-c388cdac8a", "pod":"calico-apiserver-7dd664d894-tvxc5", "timestamp":"2025-04-30 00:39:03.898191009 +0000 UTC"}, Hostname:"ci-4081.3.3-a-c388cdac8a", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.943 [INFO][5219] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.943 [INFO][5219] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.943 [INFO][5219] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.3-a-c388cdac8a' Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.946 [INFO][5219] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.951 [INFO][5219] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.976 [INFO][5219] ipam/ipam.go 489: Trying affinity for 192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.981 [INFO][5219] ipam/ipam.go 155: Attempting to load block cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.986 [INFO][5219] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.17.128/26 host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.986 [INFO][5219] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.17.128/26 handle="k8s-pod-network.42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:03.992 [INFO][5219] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:04.002 [INFO][5219] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.17.128/26 handle="k8s-pod-network.42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:04.020 [INFO][5219] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.17.134/26] block=192.168.17.128/26 handle="k8s-pod-network.42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:04.020 [INFO][5219] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.17.134/26] handle="k8s-pod-network.42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" host="ci-4081.3.3-a-c388cdac8a" Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:04.020 [INFO][5219] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:04.058163 containerd[1704]: 2025-04-30 00:39:04.020 [INFO][5219] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.17.134/26] IPv6=[] ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" HandleID="k8s-pod-network.42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:04.058810 containerd[1704]: 2025-04-30 00:39:04.024 [INFO][5197] cni-plugin/k8s.go 386: Populated endpoint ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-tvxc5" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0", GenerateName:"calico-apiserver-7dd664d894-", Namespace:"calico-apiserver", SelfLink:"", UID:"f43c40a3-2edd-4744-8016-2953f9124512", ResourceVersion:"836", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7dd664d894", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"", Pod:"calico-apiserver-7dd664d894-tvxc5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.17.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia26b2b102cd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:04.058810 containerd[1704]: 2025-04-30 00:39:04.024 [INFO][5197] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.17.134/32] ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-tvxc5" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:04.058810 containerd[1704]: 2025-04-30 00:39:04.024 [INFO][5197] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia26b2b102cd ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-tvxc5" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:04.058810 containerd[1704]: 2025-04-30 00:39:04.029 [INFO][5197] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-tvxc5" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:04.058810 containerd[1704]: 2025-04-30 00:39:04.029 [INFO][5197] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-tvxc5" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0", GenerateName:"calico-apiserver-7dd664d894-", Namespace:"calico-apiserver", SelfLink:"", UID:"f43c40a3-2edd-4744-8016-2953f9124512", ResourceVersion:"836", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7dd664d894", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae", Pod:"calico-apiserver-7dd664d894-tvxc5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.17.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia26b2b102cd", MAC:"62:b4:a5:4a:7c:5d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:04.058810 containerd[1704]: 2025-04-30 00:39:04.052 [INFO][5197] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae" Namespace="calico-apiserver" Pod="calico-apiserver-7dd664d894-tvxc5" WorkloadEndpoint="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:04.213484 containerd[1704]: time="2025-04-30T00:39:04.213262986Z" level=info msg="CreateContainer within sandbox \"487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"9bcd77cf511902bbe7898d45bb87672cb601b59b61d0f326dd31417e85d57edd\"" Apr 30 00:39:04.214560 containerd[1704]: time="2025-04-30T00:39:04.214431147Z" level=info msg="StartContainer for \"9bcd77cf511902bbe7898d45bb87672cb601b59b61d0f326dd31417e85d57edd\"" Apr 30 00:39:04.235651 containerd[1704]: time="2025-04-30T00:39:04.235496967Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 00:39:04.261888 containerd[1704]: time="2025-04-30T00:39:04.235656367Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 00:39:04.261888 containerd[1704]: time="2025-04-30T00:39:04.235696527Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:39:04.261888 containerd[1704]: time="2025-04-30T00:39:04.236255608Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 00:39:04.284328 systemd[1]: Started cri-containerd-42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae.scope - libcontainer container 42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae. Apr 30 00:39:04.285552 systemd[1]: Started cri-containerd-9bcd77cf511902bbe7898d45bb87672cb601b59b61d0f326dd31417e85d57edd.scope - libcontainer container 9bcd77cf511902bbe7898d45bb87672cb601b59b61d0f326dd31417e85d57edd. Apr 30 00:39:04.366932 containerd[1704]: time="2025-04-30T00:39:04.366890131Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7dd664d894-tvxc5,Uid:f43c40a3-2edd-4744-8016-2953f9124512,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae\"" Apr 30 00:39:04.382004 containerd[1704]: time="2025-04-30T00:39:04.381959265Z" level=info msg="StartContainer for \"9bcd77cf511902bbe7898d45bb87672cb601b59b61d0f326dd31417e85d57edd\" returns successfully" Apr 30 00:39:05.122322 systemd-networkd[1329]: calide548167888: Gained IPv6LL Apr 30 00:39:05.268909 containerd[1704]: time="2025-04-30T00:39:05.268864900Z" level=info msg="StopPodSandbox for \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\"" Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.323 [WARNING][5370] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"2d0e9630-4ffc-4065-9430-75363f7abec1", ResourceVersion:"840", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a", Pod:"coredns-6f6b679f8f-bgzm6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.17.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali989970102d8", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.323 [INFO][5370] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.323 [INFO][5370] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" iface="eth0" netns="" Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.323 [INFO][5370] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.324 [INFO][5370] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.365 [INFO][5379] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" HandleID="k8s-pod-network.c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.365 [INFO][5379] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.365 [INFO][5379] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.379 [WARNING][5379] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" HandleID="k8s-pod-network.c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.380 [INFO][5379] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" HandleID="k8s-pod-network.c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.382 [INFO][5379] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:05.391833 containerd[1704]: 2025-04-30 00:39:05.385 [INFO][5370] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:39:05.391833 containerd[1704]: time="2025-04-30T00:39:05.391024496Z" level=info msg="TearDown network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\" successfully" Apr 30 00:39:05.391833 containerd[1704]: time="2025-04-30T00:39:05.391048496Z" level=info msg="StopPodSandbox for \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\" returns successfully" Apr 30 00:39:05.394082 containerd[1704]: time="2025-04-30T00:39:05.393326338Z" level=info msg="RemovePodSandbox for \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\"" Apr 30 00:39:05.394082 containerd[1704]: time="2025-04-30T00:39:05.393369858Z" level=info msg="Forcibly stopping sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\"" Apr 30 00:39:05.442383 systemd-networkd[1329]: calia26b2b102cd: Gained IPv6LL Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.439 [WARNING][5397] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"2d0e9630-4ffc-4065-9430-75363f7abec1", ResourceVersion:"840", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"3f5c69ce554d582d92a0fa4c8186014d1d11074aef00fb9b6d9a96849a8d819a", Pod:"coredns-6f6b679f8f-bgzm6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.17.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali989970102d8", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.457 [INFO][5397] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.457 [INFO][5397] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" iface="eth0" netns="" Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.457 [INFO][5397] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.457 [INFO][5397] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.478 [INFO][5404] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" HandleID="k8s-pod-network.c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.478 [INFO][5404] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.478 [INFO][5404] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.489 [WARNING][5404] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" HandleID="k8s-pod-network.c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.489 [INFO][5404] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" HandleID="k8s-pod-network.c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--bgzm6-eth0" Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.492 [INFO][5404] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:05.496405 containerd[1704]: 2025-04-30 00:39:05.494 [INFO][5397] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225" Apr 30 00:39:05.496405 containerd[1704]: time="2025-04-30T00:39:05.496315435Z" level=info msg="TearDown network for sandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\" successfully" Apr 30 00:39:05.616970 containerd[1704]: time="2025-04-30T00:39:05.616670788Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 00:39:05.616970 containerd[1704]: time="2025-04-30T00:39:05.616761348Z" level=info msg="RemovePodSandbox \"c229c62ff61bc95d87c5537d60d2a5031c19c3144848e15dc9a2945e29931225\" returns successfully" Apr 30 00:39:05.617680 containerd[1704]: time="2025-04-30T00:39:05.617543949Z" level=info msg="StopPodSandbox for \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\"" Apr 30 00:39:05.699456 systemd-networkd[1329]: calieecd7fb4aeb: Gained IPv6LL Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.661 [WARNING][5424] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0", GenerateName:"calico-kube-controllers-768b5d7cdb-", Namespace:"calico-system", SelfLink:"", UID:"8f15ee70-a765-40e4-8bb0-3123a84281b8", ResourceVersion:"846", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"768b5d7cdb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0", Pod:"calico-kube-controllers-768b5d7cdb-g4mjx", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.17.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calide548167888", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.662 [INFO][5424] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.662 [INFO][5424] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" iface="eth0" netns="" Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.662 [INFO][5424] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.662 [INFO][5424] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.685 [INFO][5431] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" HandleID="k8s-pod-network.d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.686 [INFO][5431] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.686 [INFO][5431] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.695 [WARNING][5431] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" HandleID="k8s-pod-network.d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.695 [INFO][5431] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" HandleID="k8s-pod-network.d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.697 [INFO][5431] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:06.718127 containerd[1704]: 2025-04-30 00:39:05.700 [INFO][5424] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:06.718127 containerd[1704]: time="2025-04-30T00:39:05.703158110Z" level=info msg="TearDown network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\" successfully" Apr 30 00:39:06.718127 containerd[1704]: time="2025-04-30T00:39:05.703184110Z" level=info msg="StopPodSandbox for \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\" returns successfully" Apr 30 00:39:06.718127 containerd[1704]: time="2025-04-30T00:39:05.703770510Z" level=info msg="RemovePodSandbox for \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\"" Apr 30 00:39:06.718127 containerd[1704]: time="2025-04-30T00:39:05.703797870Z" level=info msg="Forcibly stopping sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\"" Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.780 [WARNING][5451] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0", GenerateName:"calico-kube-controllers-768b5d7cdb-", Namespace:"calico-system", SelfLink:"", UID:"8f15ee70-a765-40e4-8bb0-3123a84281b8", ResourceVersion:"846", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"768b5d7cdb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0", Pod:"calico-kube-controllers-768b5d7cdb-g4mjx", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.17.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calide548167888", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.780 [INFO][5451] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.780 [INFO][5451] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" iface="eth0" netns="" Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.780 [INFO][5451] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.780 [INFO][5451] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.803 [INFO][5458] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" HandleID="k8s-pod-network.d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.803 [INFO][5458] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.804 [INFO][5458] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.812 [WARNING][5458] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" HandleID="k8s-pod-network.d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.813 [INFO][5458] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" HandleID="k8s-pod-network.d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--kube--controllers--768b5d7cdb--g4mjx-eth0" Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.814 [INFO][5458] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:06.718724 containerd[1704]: 2025-04-30 00:39:05.816 [INFO][5451] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a" Apr 30 00:39:06.718724 containerd[1704]: time="2025-04-30T00:39:05.817765457Z" level=info msg="TearDown network for sandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\" successfully" Apr 30 00:39:07.396102 containerd[1704]: time="2025-04-30T00:39:07.395928584Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 00:39:07.396102 containerd[1704]: time="2025-04-30T00:39:07.396020384Z" level=info msg="RemovePodSandbox \"d0de33def863500ed4de4f5f5aa4340f101a9ebf545394701c79ef127ed6e55a\" returns successfully" Apr 30 00:39:07.397515 containerd[1704]: time="2025-04-30T00:39:07.397477065Z" level=info msg="StopPodSandbox for \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\"" Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.471 [WARNING][5480] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0", GenerateName:"calico-apiserver-7dd664d894-", Namespace:"calico-apiserver", SelfLink:"", UID:"f43c40a3-2edd-4744-8016-2953f9124512", ResourceVersion:"857", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7dd664d894", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae", Pod:"calico-apiserver-7dd664d894-tvxc5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.17.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia26b2b102cd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.472 [INFO][5480] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.472 [INFO][5480] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" iface="eth0" netns="" Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.472 [INFO][5480] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.472 [INFO][5480] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.499 [INFO][5487] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" HandleID="k8s-pod-network.dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.499 [INFO][5487] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.499 [INFO][5487] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.509 [WARNING][5487] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" HandleID="k8s-pod-network.dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.509 [INFO][5487] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" HandleID="k8s-pod-network.dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.510 [INFO][5487] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:07.514643 containerd[1704]: 2025-04-30 00:39:07.512 [INFO][5480] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:07.515496 containerd[1704]: time="2025-04-30T00:39:07.514823816Z" level=info msg="TearDown network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\" successfully" Apr 30 00:39:07.515496 containerd[1704]: time="2025-04-30T00:39:07.514852896Z" level=info msg="StopPodSandbox for \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\" returns successfully" Apr 30 00:39:07.515902 containerd[1704]: time="2025-04-30T00:39:07.515724737Z" level=info msg="RemovePodSandbox for \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\"" Apr 30 00:39:07.515902 containerd[1704]: time="2025-04-30T00:39:07.515763337Z" level=info msg="Forcibly stopping sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\"" Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.575 [WARNING][5505] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0", GenerateName:"calico-apiserver-7dd664d894-", Namespace:"calico-apiserver", SelfLink:"", UID:"f43c40a3-2edd-4744-8016-2953f9124512", ResourceVersion:"857", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7dd664d894", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae", Pod:"calico-apiserver-7dd664d894-tvxc5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.17.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calia26b2b102cd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.575 [INFO][5505] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.575 [INFO][5505] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" iface="eth0" netns="" Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.575 [INFO][5505] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.575 [INFO][5505] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.604 [INFO][5513] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" HandleID="k8s-pod-network.dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.604 [INFO][5513] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.604 [INFO][5513] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.617 [WARNING][5513] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" HandleID="k8s-pod-network.dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.617 [INFO][5513] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" HandleID="k8s-pod-network.dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--tvxc5-eth0" Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.620 [INFO][5513] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:07.623815 containerd[1704]: 2025-04-30 00:39:07.621 [INFO][5505] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2" Apr 30 00:39:07.624502 containerd[1704]: time="2025-04-30T00:39:07.624190870Z" level=info msg="TearDown network for sandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\" successfully" Apr 30 00:39:07.642809 containerd[1704]: time="2025-04-30T00:39:07.642741005Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 00:39:07.642976 containerd[1704]: time="2025-04-30T00:39:07.642822165Z" level=info msg="RemovePodSandbox \"dba4a32ad75efd4c9c857dc843fd7a309c5ac5ccd35502a08a20847593d85ee2\" returns successfully" Apr 30 00:39:07.643350 containerd[1704]: time="2025-04-30T00:39:07.643322366Z" level=info msg="StopPodSandbox for \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\"" Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.704 [WARNING][5531] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0", GenerateName:"calico-apiserver-7dd664d894-", Namespace:"calico-apiserver", SelfLink:"", UID:"1be863bb-f182-4a93-978b-d215d3fc0c5d", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7dd664d894", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858", Pod:"calico-apiserver-7dd664d894-lcc4f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.17.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieecd7fb4aeb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.704 [INFO][5531] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.704 [INFO][5531] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" iface="eth0" netns="" Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.704 [INFO][5531] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.704 [INFO][5531] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.752 [INFO][5538] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" HandleID="k8s-pod-network.2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.754 [INFO][5538] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.754 [INFO][5538] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.764 [WARNING][5538] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" HandleID="k8s-pod-network.2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.764 [INFO][5538] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" HandleID="k8s-pod-network.2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.767 [INFO][5538] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:07.771932 containerd[1704]: 2025-04-30 00:39:07.769 [INFO][5531] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:07.771932 containerd[1704]: time="2025-04-30T00:39:07.771653389Z" level=info msg="TearDown network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\" successfully" Apr 30 00:39:07.771932 containerd[1704]: time="2025-04-30T00:39:07.771680869Z" level=info msg="StopPodSandbox for \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\" returns successfully" Apr 30 00:39:07.773398 containerd[1704]: time="2025-04-30T00:39:07.773348711Z" level=info msg="RemovePodSandbox for \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\"" Apr 30 00:39:07.773398 containerd[1704]: time="2025-04-30T00:39:07.773398231Z" level=info msg="Forcibly stopping sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\"" Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.827 [WARNING][5556] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0", GenerateName:"calico-apiserver-7dd664d894-", Namespace:"calico-apiserver", SelfLink:"", UID:"1be863bb-f182-4a93-978b-d215d3fc0c5d", ResourceVersion:"849", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7dd664d894", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858", Pod:"calico-apiserver-7dd664d894-lcc4f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.17.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calieecd7fb4aeb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.827 [INFO][5556] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.827 [INFO][5556] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" iface="eth0" netns="" Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.827 [INFO][5556] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.827 [INFO][5556] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.855 [INFO][5563] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" HandleID="k8s-pod-network.2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.855 [INFO][5563] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.855 [INFO][5563] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.865 [WARNING][5563] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" HandleID="k8s-pod-network.2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.865 [INFO][5563] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" HandleID="k8s-pod-network.2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Workload="ci--4081.3.3--a--c388cdac8a-k8s-calico--apiserver--7dd664d894--lcc4f-eth0" Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.868 [INFO][5563] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:07.872943 containerd[1704]: 2025-04-30 00:39:07.870 [INFO][5556] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a" Apr 30 00:39:07.872943 containerd[1704]: time="2025-04-30T00:39:07.872905911Z" level=info msg="TearDown network for sandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\" successfully" Apr 30 00:39:07.888303 containerd[1704]: time="2025-04-30T00:39:07.887739643Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 00:39:07.888303 containerd[1704]: time="2025-04-30T00:39:07.887922643Z" level=info msg="RemovePodSandbox \"2169f22a12700e0722baad2460c18af854ff097213c58872af18984d7adde30a\" returns successfully" Apr 30 00:39:07.889572 containerd[1704]: time="2025-04-30T00:39:07.889538964Z" level=info msg="StopPodSandbox for \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\"" Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:07.962 [WARNING][5582] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"f93f7ee2-841a-4f1b-a68e-709e05dec366", ResourceVersion:"825", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec", Pod:"coredns-6f6b679f8f-7z7rd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.17.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7b9e637ad69", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:07.963 [INFO][5582] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:07.963 [INFO][5582] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" iface="eth0" netns="" Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:07.963 [INFO][5582] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:07.963 [INFO][5582] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:07.992 [INFO][5589] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" HandleID="k8s-pod-network.6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:07.992 [INFO][5589] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:07.992 [INFO][5589] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:08.014 [WARNING][5589] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" HandleID="k8s-pod-network.6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:08.014 [INFO][5589] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" HandleID="k8s-pod-network.6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:08.018 [INFO][5589] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:08.025886 containerd[1704]: 2025-04-30 00:39:08.023 [INFO][5582] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:39:08.025886 containerd[1704]: time="2025-04-30T00:39:08.025676514Z" level=info msg="TearDown network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\" successfully" Apr 30 00:39:08.025886 containerd[1704]: time="2025-04-30T00:39:08.025704994Z" level=info msg="StopPodSandbox for \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\" returns successfully" Apr 30 00:39:08.027312 containerd[1704]: time="2025-04-30T00:39:08.027268356Z" level=info msg="RemovePodSandbox for \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\"" Apr 30 00:39:08.027312 containerd[1704]: time="2025-04-30T00:39:08.027314356Z" level=info msg="Forcibly stopping sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\"" Apr 30 00:39:08.110762 containerd[1704]: time="2025-04-30T00:39:08.109977222Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:08.121849 containerd[1704]: time="2025-04-30T00:39:08.121768112Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.3: active requests=0, bytes read=32554116" Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.080 [WARNING][5607] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"f93f7ee2-841a-4f1b-a68e-709e05dec366", ResourceVersion:"825", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"dd9b531ee74048943d0cd248a0d4fdfed38fd58075eea47ce0170b5391b272ec", Pod:"coredns-6f6b679f8f-7z7rd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.17.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7b9e637ad69", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.080 [INFO][5607] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.080 [INFO][5607] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" iface="eth0" netns="" Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.080 [INFO][5607] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.080 [INFO][5607] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.104 [INFO][5614] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" HandleID="k8s-pod-network.6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.104 [INFO][5614] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.104 [INFO][5614] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.117 [WARNING][5614] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" HandleID="k8s-pod-network.6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.117 [INFO][5614] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" HandleID="k8s-pod-network.6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Workload="ci--4081.3.3--a--c388cdac8a-k8s-coredns--6f6b679f8f--7z7rd-eth0" Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.123 [INFO][5614] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:08.126378 containerd[1704]: 2025-04-30 00:39:08.125 [INFO][5607] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581" Apr 30 00:39:08.127484 containerd[1704]: time="2025-04-30T00:39:08.127436236Z" level=info msg="TearDown network for sandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\" successfully" Apr 30 00:39:08.134361 containerd[1704]: time="2025-04-30T00:39:08.134258162Z" level=info msg="ImageCreate event name:\"sha256:ec7c64189a2fd01b24b044fea1840d441e9884a0df32c2e9d6982cfbbea1f814\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:08.154125 containerd[1704]: time="2025-04-30T00:39:08.153025737Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 00:39:08.154125 containerd[1704]: time="2025-04-30T00:39:08.153124457Z" level=info msg="RemovePodSandbox \"6bd54467b7d01b03f5065df885bb1960281ad22d7820874361092798510f1581\" returns successfully" Apr 30 00:39:08.154125 containerd[1704]: time="2025-04-30T00:39:08.153593537Z" level=info msg="StopPodSandbox for \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\"" Apr 30 00:39:08.161982 containerd[1704]: time="2025-04-30T00:39:08.161913544Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:08.167370 containerd[1704]: time="2025-04-30T00:39:08.167319428Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" with image id \"sha256:ec7c64189a2fd01b24b044fea1840d441e9884a0df32c2e9d6982cfbbea1f814\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\", size \"33923266\" in 4.304632452s" Apr 30 00:39:08.167679 containerd[1704]: time="2025-04-30T00:39:08.167579909Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" returns image reference \"sha256:ec7c64189a2fd01b24b044fea1840d441e9884a0df32c2e9d6982cfbbea1f814\"" Apr 30 00:39:08.171753 containerd[1704]: time="2025-04-30T00:39:08.171606472Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" Apr 30 00:39:08.220517 containerd[1704]: time="2025-04-30T00:39:08.220376311Z" level=info msg="CreateContainer within sandbox \"cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Apr 30 00:39:08.278799 containerd[1704]: time="2025-04-30T00:39:08.278578238Z" level=info msg="CreateContainer within sandbox \"cf942938d4933f013ea9d583cb991464ada3abea006d4b29f76cfa4b770f6de0\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"b85e5b5d1134ecb62f92de9018accd3c970f935b7b9a953efcae2e6a6e306963\"" Apr 30 00:39:08.281102 containerd[1704]: time="2025-04-30T00:39:08.281006400Z" level=info msg="StartContainer for \"b85e5b5d1134ecb62f92de9018accd3c970f935b7b9a953efcae2e6a6e306963\"" Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.256 [WARNING][5632] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"633a0f7e-ae5b-47ef-baa9-358c2fe3b117", ResourceVersion:"788", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5bcd8f69", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c", Pod:"csi-node-driver-shvzf", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.17.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calicdda35c3166", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.256 [INFO][5632] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.256 [INFO][5632] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" iface="eth0" netns="" Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.256 [INFO][5632] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.256 [INFO][5632] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.285 [INFO][5641] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" HandleID="k8s-pod-network.9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.286 [INFO][5641] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.286 [INFO][5641] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.313 [WARNING][5641] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" HandleID="k8s-pod-network.9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.313 [INFO][5641] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" HandleID="k8s-pod-network.9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.316 [INFO][5641] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:08.322241 containerd[1704]: 2025-04-30 00:39:08.319 [INFO][5632] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:39:08.323913 containerd[1704]: time="2025-04-30T00:39:08.322329673Z" level=info msg="TearDown network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\" successfully" Apr 30 00:39:08.323913 containerd[1704]: time="2025-04-30T00:39:08.322372473Z" level=info msg="StopPodSandbox for \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\" returns successfully" Apr 30 00:39:08.323913 containerd[1704]: time="2025-04-30T00:39:08.322924314Z" level=info msg="RemovePodSandbox for \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\"" Apr 30 00:39:08.323913 containerd[1704]: time="2025-04-30T00:39:08.322955194Z" level=info msg="Forcibly stopping sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\"" Apr 30 00:39:08.323345 systemd[1]: Started cri-containerd-b85e5b5d1134ecb62f92de9018accd3c970f935b7b9a953efcae2e6a6e306963.scope - libcontainer container b85e5b5d1134ecb62f92de9018accd3c970f935b7b9a953efcae2e6a6e306963. Apr 30 00:39:08.393677 containerd[1704]: time="2025-04-30T00:39:08.393439851Z" level=info msg="StartContainer for \"b85e5b5d1134ecb62f92de9018accd3c970f935b7b9a953efcae2e6a6e306963\" returns successfully" Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.387 [WARNING][5682] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"633a0f7e-ae5b-47ef-baa9-358c2fe3b117", ResourceVersion:"788", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 0, 38, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5bcd8f69", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.3-a-c388cdac8a", ContainerID:"487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c", Pod:"csi-node-driver-shvzf", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.17.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calicdda35c3166", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.391 [INFO][5682] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.391 [INFO][5682] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" iface="eth0" netns="" Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.391 [INFO][5682] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.391 [INFO][5682] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.420 [INFO][5699] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" HandleID="k8s-pod-network.9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.420 [INFO][5699] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.420 [INFO][5699] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.435 [WARNING][5699] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" HandleID="k8s-pod-network.9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.435 [INFO][5699] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" HandleID="k8s-pod-network.9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Workload="ci--4081.3.3--a--c388cdac8a-k8s-csi--node--driver--shvzf-eth0" Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.437 [INFO][5699] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 00:39:08.441688 containerd[1704]: 2025-04-30 00:39:08.440 [INFO][5682] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0" Apr 30 00:39:08.442346 containerd[1704]: time="2025-04-30T00:39:08.441762050Z" level=info msg="TearDown network for sandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\" successfully" Apr 30 00:39:08.479115 containerd[1704]: time="2025-04-30T00:39:08.478937760Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 00:39:08.479115 containerd[1704]: time="2025-04-30T00:39:08.479035280Z" level=info msg="RemovePodSandbox \"9bd03a1799fd5f92689377898c28fec5d7871cf57e3975257bbcf7e1435b62a0\" returns successfully" Apr 30 00:39:09.485544 kubelet[3195]: I0430 00:39:09.485441 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-768b5d7cdb-g4mjx" podStartSLOduration=46.178269076 podStartE2EDuration="50.485419651s" podCreationTimestamp="2025-04-30 00:38:19 +0000 UTC" firstStartedPulling="2025-04-30 00:39:03.862159575 +0000 UTC m=+58.697385703" lastFinishedPulling="2025-04-30 00:39:08.16931015 +0000 UTC m=+63.004536278" observedRunningTime="2025-04-30 00:39:09.400284463 +0000 UTC m=+64.235510631" watchObservedRunningTime="2025-04-30 00:39:09.485419651 +0000 UTC m=+64.320645819" Apr 30 00:39:11.270484 containerd[1704]: time="2025-04-30T00:39:11.268642889Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:11.272270 containerd[1704]: time="2025-04-30T00:39:11.272042892Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=40247603" Apr 30 00:39:11.282005 containerd[1704]: time="2025-04-30T00:39:11.281909260Z" level=info msg="ImageCreate event name:\"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:11.288420 containerd[1704]: time="2025-04-30T00:39:11.288351345Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:11.290338 containerd[1704]: time="2025-04-30T00:39:11.290147587Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"41616801\" in 3.118493035s" Apr 30 00:39:11.290338 containerd[1704]: time="2025-04-30T00:39:11.290205907Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\"" Apr 30 00:39:11.292335 containerd[1704]: time="2025-04-30T00:39:11.292151508Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" Apr 30 00:39:11.295596 containerd[1704]: time="2025-04-30T00:39:11.295536911Z" level=info msg="CreateContainer within sandbox \"d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 30 00:39:11.346866 containerd[1704]: time="2025-04-30T00:39:11.346710992Z" level=info msg="CreateContainer within sandbox \"d8132141533f9a77e30cea6308b33bb13641a7993d56c20056086cc366ad7858\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"91af50d25f76bf61b6124bd280fb102f34455e8194511650819bf8de5659c6f7\"" Apr 30 00:39:11.349166 containerd[1704]: time="2025-04-30T00:39:11.348350874Z" level=info msg="StartContainer for \"91af50d25f76bf61b6124bd280fb102f34455e8194511650819bf8de5659c6f7\"" Apr 30 00:39:11.399020 systemd[1]: Started cri-containerd-91af50d25f76bf61b6124bd280fb102f34455e8194511650819bf8de5659c6f7.scope - libcontainer container 91af50d25f76bf61b6124bd280fb102f34455e8194511650819bf8de5659c6f7. Apr 30 00:39:11.451615 containerd[1704]: time="2025-04-30T00:39:11.451402837Z" level=info msg="StartContainer for \"91af50d25f76bf61b6124bd280fb102f34455e8194511650819bf8de5659c6f7\" returns successfully" Apr 30 00:39:11.655359 containerd[1704]: time="2025-04-30T00:39:11.655198241Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:11.661444 containerd[1704]: time="2025-04-30T00:39:11.658914924Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=77" Apr 30 00:39:11.662634 containerd[1704]: time="2025-04-30T00:39:11.662584247Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"41616801\" in 370.369459ms" Apr 30 00:39:11.662805 containerd[1704]: time="2025-04-30T00:39:11.662787207Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\"" Apr 30 00:39:11.665760 containerd[1704]: time="2025-04-30T00:39:11.665715010Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\"" Apr 30 00:39:11.666865 containerd[1704]: time="2025-04-30T00:39:11.666820690Z" level=info msg="CreateContainer within sandbox \"42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 30 00:39:11.739798 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount173224039.mount: Deactivated successfully. Apr 30 00:39:11.755772 containerd[1704]: time="2025-04-30T00:39:11.755683082Z" level=info msg="CreateContainer within sandbox \"42fa6a14ae5b1778266c80acb5e4de6702258298a495df7894bbaf5b4e1d22ae\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"0ab5bedb33d560c161b71a8bdbe0cabbfa134ff630a567089d57d7b815f121a0\"" Apr 30 00:39:11.756756 containerd[1704]: time="2025-04-30T00:39:11.756573523Z" level=info msg="StartContainer for \"0ab5bedb33d560c161b71a8bdbe0cabbfa134ff630a567089d57d7b815f121a0\"" Apr 30 00:39:11.803291 systemd[1]: Started cri-containerd-0ab5bedb33d560c161b71a8bdbe0cabbfa134ff630a567089d57d7b815f121a0.scope - libcontainer container 0ab5bedb33d560c161b71a8bdbe0cabbfa134ff630a567089d57d7b815f121a0. Apr 30 00:39:11.849369 containerd[1704]: time="2025-04-30T00:39:11.849315678Z" level=info msg="StartContainer for \"0ab5bedb33d560c161b71a8bdbe0cabbfa134ff630a567089d57d7b815f121a0\" returns successfully" Apr 30 00:39:12.493021 kubelet[3195]: I0430 00:39:12.492940 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7dd664d894-lcc4f" podStartSLOduration=47.205339478 podStartE2EDuration="54.492921517s" podCreationTimestamp="2025-04-30 00:38:18 +0000 UTC" firstStartedPulling="2025-04-30 00:39:04.004292189 +0000 UTC m=+58.839518357" lastFinishedPulling="2025-04-30 00:39:11.291874148 +0000 UTC m=+66.127100396" observedRunningTime="2025-04-30 00:39:12.481527307 +0000 UTC m=+67.316753475" watchObservedRunningTime="2025-04-30 00:39:12.492921517 +0000 UTC m=+67.328147685" Apr 30 00:39:12.493465 kubelet[3195]: I0430 00:39:12.493048 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7dd664d894-tvxc5" podStartSLOduration=47.199964244 podStartE2EDuration="54.493042677s" podCreationTimestamp="2025-04-30 00:38:18 +0000 UTC" firstStartedPulling="2025-04-30 00:39:04.371004055 +0000 UTC m=+59.206230223" lastFinishedPulling="2025-04-30 00:39:11.664082488 +0000 UTC m=+66.499308656" observedRunningTime="2025-04-30 00:39:12.440414554 +0000 UTC m=+67.275640722" watchObservedRunningTime="2025-04-30 00:39:12.493042677 +0000 UTC m=+67.328268845" Apr 30 00:39:13.405674 containerd[1704]: time="2025-04-30T00:39:13.405588013Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:13.412237 containerd[1704]: time="2025-04-30T00:39:13.412182018Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3: active requests=0, bytes read=13124299" Apr 30 00:39:13.417792 containerd[1704]: time="2025-04-30T00:39:13.417721462Z" level=info msg="ImageCreate event name:\"sha256:a91b1f00752edc175f270a01b33683fa80818734aa2274388785eaf3364315dc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:13.432618 containerd[1704]: time="2025-04-30T00:39:13.432541154Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 00:39:13.433381 containerd[1704]: time="2025-04-30T00:39:13.433339995Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" with image id \"sha256:a91b1f00752edc175f270a01b33683fa80818734aa2274388785eaf3364315dc\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\", size \"14493433\" in 1.767232905s" Apr 30 00:39:13.433381 containerd[1704]: time="2025-04-30T00:39:13.433381435Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" returns image reference \"sha256:a91b1f00752edc175f270a01b33683fa80818734aa2274388785eaf3364315dc\"" Apr 30 00:39:13.441649 containerd[1704]: time="2025-04-30T00:39:13.441589282Z" level=info msg="CreateContainer within sandbox \"487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Apr 30 00:39:13.514400 containerd[1704]: time="2025-04-30T00:39:13.514176140Z" level=info msg="CreateContainer within sandbox \"487197b72cb3c90d9783549ec5ce4b16727123e758322b10539d0b6d4f7ff78c\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"36d13c0659059fcc8f9be67c0941d77253bf42893485142f71947b42bc637e16\"" Apr 30 00:39:13.516474 containerd[1704]: time="2025-04-30T00:39:13.515454781Z" level=info msg="StartContainer for \"36d13c0659059fcc8f9be67c0941d77253bf42893485142f71947b42bc637e16\"" Apr 30 00:39:13.582315 systemd[1]: Started cri-containerd-36d13c0659059fcc8f9be67c0941d77253bf42893485142f71947b42bc637e16.scope - libcontainer container 36d13c0659059fcc8f9be67c0941d77253bf42893485142f71947b42bc637e16. Apr 30 00:39:13.640928 containerd[1704]: time="2025-04-30T00:39:13.640520122Z" level=info msg="StartContainer for \"36d13c0659059fcc8f9be67c0941d77253bf42893485142f71947b42bc637e16\" returns successfully" Apr 30 00:39:14.438902 kubelet[3195]: I0430 00:39:14.438852 3195 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Apr 30 00:39:14.438902 kubelet[3195]: I0430 00:39:14.438918 3195 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Apr 30 00:39:14.448435 kubelet[3195]: I0430 00:39:14.447983 3195 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-shvzf" podStartSLOduration=39.097186134 podStartE2EDuration="55.447961693s" podCreationTimestamp="2025-04-30 00:38:19 +0000 UTC" firstStartedPulling="2025-04-30 00:38:57.08750564 +0000 UTC m=+51.922731808" lastFinishedPulling="2025-04-30 00:39:13.438281199 +0000 UTC m=+68.273507367" observedRunningTime="2025-04-30 00:39:14.44386777 +0000 UTC m=+69.279093938" watchObservedRunningTime="2025-04-30 00:39:14.447961693 +0000 UTC m=+69.283187861" Apr 30 00:39:44.643248 systemd[1]: run-containerd-runc-k8s.io-b85e5b5d1134ecb62f92de9018accd3c970f935b7b9a953efcae2e6a6e306963-runc.fHUibL.mount: Deactivated successfully. Apr 30 00:39:56.915141 systemd[1]: run-containerd-runc-k8s.io-cdcaefa1cf2354133d788af5005110c406247f63aab7d224d35aff04db810fdd-runc.GtyD4L.mount: Deactivated successfully. Apr 30 00:40:45.456391 systemd[1]: Started sshd@7-10.200.20.12:22-10.200.16.10:34850.service - OpenSSH per-connection server daemon (10.200.16.10:34850). Apr 30 00:40:45.908460 sshd[6084]: Accepted publickey for core from 10.200.16.10 port 34850 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:40:45.912422 sshd[6084]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:40:45.923143 systemd-logind[1668]: New session 10 of user core. Apr 30 00:40:45.928310 systemd[1]: Started session-10.scope - Session 10 of User core. Apr 30 00:40:46.341268 sshd[6084]: pam_unix(sshd:session): session closed for user core Apr 30 00:40:46.345712 systemd[1]: sshd@7-10.200.20.12:22-10.200.16.10:34850.service: Deactivated successfully. Apr 30 00:40:46.350355 systemd[1]: session-10.scope: Deactivated successfully. Apr 30 00:40:46.354566 systemd-logind[1668]: Session 10 logged out. Waiting for processes to exit. Apr 30 00:40:46.357378 systemd-logind[1668]: Removed session 10. Apr 30 00:40:51.423351 systemd[1]: Started sshd@8-10.200.20.12:22-10.200.16.10:36070.service - OpenSSH per-connection server daemon (10.200.16.10:36070). Apr 30 00:40:51.837960 sshd[6099]: Accepted publickey for core from 10.200.16.10 port 36070 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:40:51.839367 sshd[6099]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:40:51.843948 systemd-logind[1668]: New session 11 of user core. Apr 30 00:40:51.849292 systemd[1]: Started session-11.scope - Session 11 of User core. Apr 30 00:40:52.208725 sshd[6099]: pam_unix(sshd:session): session closed for user core Apr 30 00:40:52.212791 systemd-logind[1668]: Session 11 logged out. Waiting for processes to exit. Apr 30 00:40:52.213437 systemd[1]: sshd@8-10.200.20.12:22-10.200.16.10:36070.service: Deactivated successfully. Apr 30 00:40:52.215624 systemd[1]: session-11.scope: Deactivated successfully. Apr 30 00:40:52.217980 systemd-logind[1668]: Removed session 11. Apr 30 00:40:57.306362 systemd[1]: Started sshd@9-10.200.20.12:22-10.200.16.10:36078.service - OpenSSH per-connection server daemon (10.200.16.10:36078). Apr 30 00:40:57.755375 sshd[6135]: Accepted publickey for core from 10.200.16.10 port 36078 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:40:57.756915 sshd[6135]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:40:57.761649 systemd-logind[1668]: New session 12 of user core. Apr 30 00:40:57.766471 systemd[1]: Started session-12.scope - Session 12 of User core. Apr 30 00:40:58.156122 sshd[6135]: pam_unix(sshd:session): session closed for user core Apr 30 00:40:58.159450 systemd-logind[1668]: Session 12 logged out. Waiting for processes to exit. Apr 30 00:40:58.160570 systemd[1]: sshd@9-10.200.20.12:22-10.200.16.10:36078.service: Deactivated successfully. Apr 30 00:40:58.163725 systemd[1]: session-12.scope: Deactivated successfully. Apr 30 00:40:58.166785 systemd-logind[1668]: Removed session 12. Apr 30 00:41:03.241411 systemd[1]: Started sshd@10-10.200.20.12:22-10.200.16.10:43710.service - OpenSSH per-connection server daemon (10.200.16.10:43710). Apr 30 00:41:03.689975 sshd[6149]: Accepted publickey for core from 10.200.16.10 port 43710 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:03.695920 sshd[6149]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:03.700718 systemd-logind[1668]: New session 13 of user core. Apr 30 00:41:03.707530 systemd[1]: Started session-13.scope - Session 13 of User core. Apr 30 00:41:04.081351 sshd[6149]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:04.086890 systemd-logind[1668]: Session 13 logged out. Waiting for processes to exit. Apr 30 00:41:04.087549 systemd[1]: sshd@10-10.200.20.12:22-10.200.16.10:43710.service: Deactivated successfully. Apr 30 00:41:04.090850 systemd[1]: session-13.scope: Deactivated successfully. Apr 30 00:41:04.092590 systemd-logind[1668]: Removed session 13. Apr 30 00:41:04.161405 systemd[1]: Started sshd@11-10.200.20.12:22-10.200.16.10:43716.service - OpenSSH per-connection server daemon (10.200.16.10:43716). Apr 30 00:41:04.571613 sshd[6181]: Accepted publickey for core from 10.200.16.10 port 43716 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:04.573254 sshd[6181]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:04.578293 systemd-logind[1668]: New session 14 of user core. Apr 30 00:41:04.581288 systemd[1]: Started session-14.scope - Session 14 of User core. Apr 30 00:41:05.012654 sshd[6181]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:05.018751 systemd[1]: sshd@11-10.200.20.12:22-10.200.16.10:43716.service: Deactivated successfully. Apr 30 00:41:05.024969 systemd[1]: session-14.scope: Deactivated successfully. Apr 30 00:41:05.026571 systemd-logind[1668]: Session 14 logged out. Waiting for processes to exit. Apr 30 00:41:05.029142 systemd-logind[1668]: Removed session 14. Apr 30 00:41:05.101405 systemd[1]: Started sshd@12-10.200.20.12:22-10.200.16.10:43730.service - OpenSSH per-connection server daemon (10.200.16.10:43730). Apr 30 00:41:05.523412 sshd[6195]: Accepted publickey for core from 10.200.16.10 port 43730 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:05.524940 sshd[6195]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:05.529804 systemd-logind[1668]: New session 15 of user core. Apr 30 00:41:05.535312 systemd[1]: Started session-15.scope - Session 15 of User core. Apr 30 00:41:05.895200 sshd[6195]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:05.899258 systemd-logind[1668]: Session 15 logged out. Waiting for processes to exit. Apr 30 00:41:05.900049 systemd[1]: sshd@12-10.200.20.12:22-10.200.16.10:43730.service: Deactivated successfully. Apr 30 00:41:05.903483 systemd[1]: session-15.scope: Deactivated successfully. Apr 30 00:41:05.905104 systemd-logind[1668]: Removed session 15. Apr 30 00:41:10.977234 systemd[1]: Started sshd@13-10.200.20.12:22-10.200.16.10:60566.service - OpenSSH per-connection server daemon (10.200.16.10:60566). Apr 30 00:41:11.434660 sshd[6213]: Accepted publickey for core from 10.200.16.10 port 60566 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:11.436169 sshd[6213]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:11.440908 systemd-logind[1668]: New session 16 of user core. Apr 30 00:41:11.446260 systemd[1]: Started session-16.scope - Session 16 of User core. Apr 30 00:41:11.824592 sshd[6213]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:11.828096 systemd-logind[1668]: Session 16 logged out. Waiting for processes to exit. Apr 30 00:41:11.828281 systemd[1]: sshd@13-10.200.20.12:22-10.200.16.10:60566.service: Deactivated successfully. Apr 30 00:41:11.830241 systemd[1]: session-16.scope: Deactivated successfully. Apr 30 00:41:11.833588 systemd-logind[1668]: Removed session 16. Apr 30 00:41:11.901279 systemd[1]: Started sshd@14-10.200.20.12:22-10.200.16.10:60578.service - OpenSSH per-connection server daemon (10.200.16.10:60578). Apr 30 00:41:12.321480 sshd[6225]: Accepted publickey for core from 10.200.16.10 port 60578 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:12.322993 sshd[6225]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:12.327185 systemd-logind[1668]: New session 17 of user core. Apr 30 00:41:12.331416 systemd[1]: Started session-17.scope - Session 17 of User core. Apr 30 00:41:12.791730 sshd[6225]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:12.796110 systemd[1]: sshd@14-10.200.20.12:22-10.200.16.10:60578.service: Deactivated successfully. Apr 30 00:41:12.798348 systemd[1]: session-17.scope: Deactivated successfully. Apr 30 00:41:12.799208 systemd-logind[1668]: Session 17 logged out. Waiting for processes to exit. Apr 30 00:41:12.800239 systemd-logind[1668]: Removed session 17. Apr 30 00:41:12.876385 systemd[1]: Started sshd@15-10.200.20.12:22-10.200.16.10:60586.service - OpenSSH per-connection server daemon (10.200.16.10:60586). Apr 30 00:41:13.319884 sshd[6235]: Accepted publickey for core from 10.200.16.10 port 60586 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:13.321390 sshd[6235]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:13.326304 systemd-logind[1668]: New session 18 of user core. Apr 30 00:41:13.333256 systemd[1]: Started session-18.scope - Session 18 of User core. Apr 30 00:41:15.426192 sshd[6235]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:15.432451 systemd[1]: sshd@15-10.200.20.12:22-10.200.16.10:60586.service: Deactivated successfully. Apr 30 00:41:15.436037 systemd[1]: session-18.scope: Deactivated successfully. Apr 30 00:41:15.439666 systemd-logind[1668]: Session 18 logged out. Waiting for processes to exit. Apr 30 00:41:15.442155 systemd-logind[1668]: Removed session 18. Apr 30 00:41:15.513398 systemd[1]: Started sshd@16-10.200.20.12:22-10.200.16.10:60596.service - OpenSSH per-connection server daemon (10.200.16.10:60596). Apr 30 00:41:15.968974 sshd[6272]: Accepted publickey for core from 10.200.16.10 port 60596 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:15.970403 sshd[6272]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:15.974991 systemd-logind[1668]: New session 19 of user core. Apr 30 00:41:15.979321 systemd[1]: Started session-19.scope - Session 19 of User core. Apr 30 00:41:16.544310 sshd[6272]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:16.549737 systemd-logind[1668]: Session 19 logged out. Waiting for processes to exit. Apr 30 00:41:16.551491 systemd[1]: sshd@16-10.200.20.12:22-10.200.16.10:60596.service: Deactivated successfully. Apr 30 00:41:16.556140 systemd[1]: session-19.scope: Deactivated successfully. Apr 30 00:41:16.559203 systemd-logind[1668]: Removed session 19. Apr 30 00:41:16.635594 systemd[1]: Started sshd@17-10.200.20.12:22-10.200.16.10:60612.service - OpenSSH per-connection server daemon (10.200.16.10:60612). Apr 30 00:41:17.052027 sshd[6283]: Accepted publickey for core from 10.200.16.10 port 60612 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:17.053500 sshd[6283]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:17.057534 systemd-logind[1668]: New session 20 of user core. Apr 30 00:41:17.067464 systemd[1]: Started session-20.scope - Session 20 of User core. Apr 30 00:41:17.421590 sshd[6283]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:17.426400 systemd[1]: sshd@17-10.200.20.12:22-10.200.16.10:60612.service: Deactivated successfully. Apr 30 00:41:17.428880 systemd[1]: session-20.scope: Deactivated successfully. Apr 30 00:41:17.430719 systemd-logind[1668]: Session 20 logged out. Waiting for processes to exit. Apr 30 00:41:17.433008 systemd-logind[1668]: Removed session 20. Apr 30 00:41:22.522396 systemd[1]: Started sshd@18-10.200.20.12:22-10.200.16.10:46986.service - OpenSSH per-connection server daemon (10.200.16.10:46986). Apr 30 00:41:22.967006 sshd[6299]: Accepted publickey for core from 10.200.16.10 port 46986 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:22.968638 sshd[6299]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:22.972856 systemd-logind[1668]: New session 21 of user core. Apr 30 00:41:22.980293 systemd[1]: Started session-21.scope - Session 21 of User core. Apr 30 00:41:23.357858 sshd[6299]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:23.362039 systemd-logind[1668]: Session 21 logged out. Waiting for processes to exit. Apr 30 00:41:23.362879 systemd[1]: sshd@18-10.200.20.12:22-10.200.16.10:46986.service: Deactivated successfully. Apr 30 00:41:23.366771 systemd[1]: session-21.scope: Deactivated successfully. Apr 30 00:41:23.367893 systemd-logind[1668]: Removed session 21. Apr 30 00:41:28.439465 systemd[1]: Started sshd@19-10.200.20.12:22-10.200.16.10:47002.service - OpenSSH per-connection server daemon (10.200.16.10:47002). Apr 30 00:41:28.893793 sshd[6335]: Accepted publickey for core from 10.200.16.10 port 47002 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:28.895520 sshd[6335]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:28.900125 systemd-logind[1668]: New session 22 of user core. Apr 30 00:41:28.903257 systemd[1]: Started session-22.scope - Session 22 of User core. Apr 30 00:41:29.280610 sshd[6335]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:29.284549 systemd[1]: sshd@19-10.200.20.12:22-10.200.16.10:47002.service: Deactivated successfully. Apr 30 00:41:29.288601 systemd[1]: session-22.scope: Deactivated successfully. Apr 30 00:41:29.291949 systemd-logind[1668]: Session 22 logged out. Waiting for processes to exit. Apr 30 00:41:29.295017 systemd-logind[1668]: Removed session 22. Apr 30 00:41:34.360740 systemd[1]: Started sshd@20-10.200.20.12:22-10.200.16.10:39714.service - OpenSSH per-connection server daemon (10.200.16.10:39714). Apr 30 00:41:34.785793 sshd[6355]: Accepted publickey for core from 10.200.16.10 port 39714 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:34.787395 sshd[6355]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:34.793614 systemd-logind[1668]: New session 23 of user core. Apr 30 00:41:34.799277 systemd[1]: Started session-23.scope - Session 23 of User core. Apr 30 00:41:35.149249 sshd[6355]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:35.153481 systemd[1]: sshd@20-10.200.20.12:22-10.200.16.10:39714.service: Deactivated successfully. Apr 30 00:41:35.156187 systemd[1]: session-23.scope: Deactivated successfully. Apr 30 00:41:35.157190 systemd-logind[1668]: Session 23 logged out. Waiting for processes to exit. Apr 30 00:41:35.158724 systemd-logind[1668]: Removed session 23. Apr 30 00:41:40.230400 systemd[1]: Started sshd@21-10.200.20.12:22-10.200.16.10:39994.service - OpenSSH per-connection server daemon (10.200.16.10:39994). Apr 30 00:41:40.643688 sshd[6371]: Accepted publickey for core from 10.200.16.10 port 39994 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:40.645324 sshd[6371]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:40.650213 systemd-logind[1668]: New session 24 of user core. Apr 30 00:41:40.660308 systemd[1]: Started session-24.scope - Session 24 of User core. Apr 30 00:41:41.011550 sshd[6371]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:41.019252 systemd[1]: sshd@21-10.200.20.12:22-10.200.16.10:39994.service: Deactivated successfully. Apr 30 00:41:41.023207 systemd[1]: session-24.scope: Deactivated successfully. Apr 30 00:41:41.025320 systemd-logind[1668]: Session 24 logged out. Waiting for processes to exit. Apr 30 00:41:41.026872 systemd-logind[1668]: Removed session 24. Apr 30 00:41:46.100413 systemd[1]: Started sshd@22-10.200.20.12:22-10.200.16.10:39998.service - OpenSSH per-connection server daemon (10.200.16.10:39998). Apr 30 00:41:46.548031 sshd[6410]: Accepted publickey for core from 10.200.16.10 port 39998 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:46.550156 sshd[6410]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:46.556269 systemd-logind[1668]: New session 25 of user core. Apr 30 00:41:46.562348 systemd[1]: Started session-25.scope - Session 25 of User core. Apr 30 00:41:46.948708 sshd[6410]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:46.952481 systemd[1]: sshd@22-10.200.20.12:22-10.200.16.10:39998.service: Deactivated successfully. Apr 30 00:41:46.955518 systemd[1]: session-25.scope: Deactivated successfully. Apr 30 00:41:46.956932 systemd-logind[1668]: Session 25 logged out. Waiting for processes to exit. Apr 30 00:41:46.958050 systemd-logind[1668]: Removed session 25. Apr 30 00:41:52.035384 systemd[1]: Started sshd@23-10.200.20.12:22-10.200.16.10:53754.service - OpenSSH per-connection server daemon (10.200.16.10:53754). Apr 30 00:41:52.446646 sshd[6423]: Accepted publickey for core from 10.200.16.10 port 53754 ssh2: RSA SHA256:ztpvO7lq7UFkG/gUNSQtdxecuZ/3hQtQILcGfuKW7pw Apr 30 00:41:52.448295 sshd[6423]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 00:41:52.453427 systemd-logind[1668]: New session 26 of user core. Apr 30 00:41:52.457289 systemd[1]: Started session-26.scope - Session 26 of User core. Apr 30 00:41:52.813240 sshd[6423]: pam_unix(sshd:session): session closed for user core Apr 30 00:41:52.819126 systemd-logind[1668]: Session 26 logged out. Waiting for processes to exit. Apr 30 00:41:52.820151 systemd[1]: sshd@23-10.200.20.12:22-10.200.16.10:53754.service: Deactivated successfully. Apr 30 00:41:52.824751 systemd[1]: session-26.scope: Deactivated successfully. Apr 30 00:41:52.826455 systemd-logind[1668]: Removed session 26.