Aug 13 00:26:46.983355 kernel: Linux version 6.12.40-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Tue Aug 12 21:42:48 -00 2025 Aug 13 00:26:46.983386 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=215bdedb8de38f6b96ec4f9db80853e25015f60454b867e319fdcb9244320a21 Aug 13 00:26:46.983396 kernel: BIOS-provided physical RAM map: Aug 13 00:26:46.983403 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Aug 13 00:26:46.983409 kernel: BIOS-e820: [mem 0x00000000000c0000-0x00000000000fffff] reserved Aug 13 00:26:46.983416 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000044fdfff] usable Aug 13 00:26:46.983424 kernel: BIOS-e820: [mem 0x00000000044fe000-0x00000000048fdfff] reserved Aug 13 00:26:46.983433 kernel: BIOS-e820: [mem 0x00000000048fe000-0x000000003ff1efff] usable Aug 13 00:26:46.983439 kernel: BIOS-e820: [mem 0x000000003ff1f000-0x000000003ffc8fff] reserved Aug 13 00:26:46.983446 kernel: BIOS-e820: [mem 0x000000003ffc9000-0x000000003fffafff] ACPI data Aug 13 00:26:46.983453 kernel: BIOS-e820: [mem 0x000000003fffb000-0x000000003fffefff] ACPI NVS Aug 13 00:26:46.983460 kernel: BIOS-e820: [mem 0x000000003ffff000-0x000000003fffffff] usable Aug 13 00:26:46.983466 kernel: BIOS-e820: [mem 0x0000000100000000-0x00000002bfffffff] usable Aug 13 00:26:46.983472 kernel: printk: legacy bootconsole [earlyser0] enabled Aug 13 00:26:46.983482 kernel: NX (Execute Disable) protection: active Aug 13 00:26:46.983489 kernel: APIC: Static calls initialized Aug 13 00:26:46.983496 kernel: efi: EFI v2.7 by Microsoft Aug 13 00:26:46.983504 kernel: efi: ACPI=0x3fffa000 ACPI 2.0=0x3fffa014 SMBIOS=0x3ff88000 SMBIOS 3.0=0x3ff86000 MEMATTR=0x3e9da518 RNG=0x3ffd2018 Aug 13 00:26:46.983511 kernel: random: crng init done Aug 13 00:26:46.983519 kernel: secureboot: Secure boot disabled Aug 13 00:26:46.983526 kernel: SMBIOS 3.1.0 present. Aug 13 00:26:46.983534 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 01/28/2025 Aug 13 00:26:46.983541 kernel: DMI: Memory slots populated: 2/2 Aug 13 00:26:46.983549 kernel: Hypervisor detected: Microsoft Hyper-V Aug 13 00:26:46.983557 kernel: Hyper-V: privilege flags low 0xae7f, high 0x3b8030, hints 0x9e4e24, misc 0xe0bed7b2 Aug 13 00:26:46.983563 kernel: Hyper-V: Nested features: 0x3e0101 Aug 13 00:26:46.983570 kernel: Hyper-V: LAPIC Timer Frequency: 0x30d40 Aug 13 00:26:46.983578 kernel: Hyper-V: Using hypercall for remote TLB flush Aug 13 00:26:46.983585 kernel: clocksource: hyperv_clocksource_tsc_page: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Aug 13 00:26:46.983593 kernel: clocksource: hyperv_clocksource_msr: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Aug 13 00:26:46.983600 kernel: tsc: Detected 2300.000 MHz processor Aug 13 00:26:46.983607 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Aug 13 00:26:46.983615 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Aug 13 00:26:46.983625 kernel: last_pfn = 0x2c0000 max_arch_pfn = 0x10000000000 Aug 13 00:26:46.983633 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Aug 13 00:26:46.983640 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Aug 13 00:26:46.983648 kernel: e820: update [mem 0x48000000-0xffffffff] usable ==> reserved Aug 13 00:26:46.983655 kernel: last_pfn = 0x40000 max_arch_pfn = 0x10000000000 Aug 13 00:26:46.983663 kernel: Using GB pages for direct mapping Aug 13 00:26:46.983671 kernel: ACPI: Early table checksum verification disabled Aug 13 00:26:46.983681 kernel: ACPI: RSDP 0x000000003FFFA014 000024 (v02 VRTUAL) Aug 13 00:26:46.983691 kernel: ACPI: XSDT 0x000000003FFF90E8 00005C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 00:26:46.983699 kernel: ACPI: FACP 0x000000003FFF8000 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 00:26:46.983707 kernel: ACPI: DSDT 0x000000003FFD6000 01E27A (v02 MSFTVM DSDT01 00000001 INTL 20230628) Aug 13 00:26:46.983714 kernel: ACPI: FACS 0x000000003FFFE000 000040 Aug 13 00:26:46.983722 kernel: ACPI: OEM0 0x000000003FFF7000 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 00:26:46.983730 kernel: ACPI: SPCR 0x000000003FFF6000 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 00:26:46.983740 kernel: ACPI: WAET 0x000000003FFF5000 000028 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 00:26:46.983748 kernel: ACPI: APIC 0x000000003FFD5000 000052 (v05 HVLITE HVLITETB 00000000 MSHV 00000000) Aug 13 00:26:46.983756 kernel: ACPI: SRAT 0x000000003FFD4000 0000A0 (v03 HVLITE HVLITETB 00000000 MSHV 00000000) Aug 13 00:26:46.983764 kernel: ACPI: BGRT 0x000000003FFD3000 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 00:26:46.983773 kernel: ACPI: Reserving FACP table memory at [mem 0x3fff8000-0x3fff8113] Aug 13 00:26:46.983781 kernel: ACPI: Reserving DSDT table memory at [mem 0x3ffd6000-0x3fff4279] Aug 13 00:26:46.983788 kernel: ACPI: Reserving FACS table memory at [mem 0x3fffe000-0x3fffe03f] Aug 13 00:26:46.983796 kernel: ACPI: Reserving OEM0 table memory at [mem 0x3fff7000-0x3fff7063] Aug 13 00:26:46.983803 kernel: ACPI: Reserving SPCR table memory at [mem 0x3fff6000-0x3fff604f] Aug 13 00:26:46.983812 kernel: ACPI: Reserving WAET table memory at [mem 0x3fff5000-0x3fff5027] Aug 13 00:26:46.983820 kernel: ACPI: Reserving APIC table memory at [mem 0x3ffd5000-0x3ffd5051] Aug 13 00:26:46.983828 kernel: ACPI: Reserving SRAT table memory at [mem 0x3ffd4000-0x3ffd409f] Aug 13 00:26:46.983835 kernel: ACPI: Reserving BGRT table memory at [mem 0x3ffd3000-0x3ffd3037] Aug 13 00:26:46.983843 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] Aug 13 00:26:46.983851 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x2bfffffff] Aug 13 00:26:46.983858 kernel: NUMA: Node 0 [mem 0x00001000-0x3fffffff] + [mem 0x100000000-0x2bfffffff] -> [mem 0x00001000-0x2bfffffff] Aug 13 00:26:46.983866 kernel: NODE_DATA(0) allocated [mem 0x2bfff8dc0-0x2bfffffff] Aug 13 00:26:46.983873 kernel: Zone ranges: Aug 13 00:26:46.983882 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Aug 13 00:26:46.983890 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Aug 13 00:26:46.983898 kernel: Normal [mem 0x0000000100000000-0x00000002bfffffff] Aug 13 00:26:46.983905 kernel: Device empty Aug 13 00:26:46.983913 kernel: Movable zone start for each node Aug 13 00:26:46.983921 kernel: Early memory node ranges Aug 13 00:26:46.983928 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Aug 13 00:26:46.983936 kernel: node 0: [mem 0x0000000000100000-0x00000000044fdfff] Aug 13 00:26:46.983943 kernel: node 0: [mem 0x00000000048fe000-0x000000003ff1efff] Aug 13 00:26:46.983952 kernel: node 0: [mem 0x000000003ffff000-0x000000003fffffff] Aug 13 00:26:46.983960 kernel: node 0: [mem 0x0000000100000000-0x00000002bfffffff] Aug 13 00:26:46.983967 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x00000002bfffffff] Aug 13 00:26:46.983975 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Aug 13 00:26:46.983982 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Aug 13 00:26:46.983990 kernel: On node 0, zone DMA32: 1024 pages in unavailable ranges Aug 13 00:26:46.983998 kernel: On node 0, zone DMA32: 224 pages in unavailable ranges Aug 13 00:26:46.984005 kernel: ACPI: PM-Timer IO Port: 0x408 Aug 13 00:26:46.984013 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Aug 13 00:26:46.984022 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Aug 13 00:26:46.984029 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Aug 13 00:26:46.984037 kernel: ACPI: SPCR: console: uart,io,0x3f8,115200 Aug 13 00:26:46.984044 kernel: TSC deadline timer available Aug 13 00:26:46.984052 kernel: CPU topo: Max. logical packages: 1 Aug 13 00:26:46.984058 kernel: CPU topo: Max. logical dies: 1 Aug 13 00:26:46.984065 kernel: CPU topo: Max. dies per package: 1 Aug 13 00:26:46.984073 kernel: CPU topo: Max. threads per core: 2 Aug 13 00:26:46.984080 kernel: CPU topo: Num. cores per package: 1 Aug 13 00:26:46.984089 kernel: CPU topo: Num. threads per package: 2 Aug 13 00:26:46.984097 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Aug 13 00:26:46.984104 kernel: [mem 0x40000000-0xffffffff] available for PCI devices Aug 13 00:26:46.984112 kernel: Booting paravirtualized kernel on Hyper-V Aug 13 00:26:46.984119 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Aug 13 00:26:46.984127 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Aug 13 00:26:46.984135 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Aug 13 00:26:46.984142 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Aug 13 00:26:46.984150 kernel: pcpu-alloc: [0] 0 1 Aug 13 00:26:46.984160 kernel: Hyper-V: PV spinlocks enabled Aug 13 00:26:46.984168 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Aug 13 00:26:46.984177 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=215bdedb8de38f6b96ec4f9db80853e25015f60454b867e319fdcb9244320a21 Aug 13 00:26:46.984186 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 00:26:46.984193 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) Aug 13 00:26:46.984201 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Aug 13 00:26:46.984208 kernel: Fallback order for Node 0: 0 Aug 13 00:26:46.984216 kernel: Built 1 zonelists, mobility grouping on. Total pages: 2095807 Aug 13 00:26:46.984226 kernel: Policy zone: Normal Aug 13 00:26:46.984233 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 00:26:46.984241 kernel: software IO TLB: area num 2. Aug 13 00:26:46.984249 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Aug 13 00:26:46.985278 kernel: ftrace: allocating 40098 entries in 157 pages Aug 13 00:26:46.985298 kernel: ftrace: allocated 157 pages with 5 groups Aug 13 00:26:46.985306 kernel: Dynamic Preempt: voluntary Aug 13 00:26:46.985314 kernel: rcu: Preemptible hierarchical RCU implementation. Aug 13 00:26:46.985324 kernel: rcu: RCU event tracing is enabled. Aug 13 00:26:46.985342 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Aug 13 00:26:46.985350 kernel: Trampoline variant of Tasks RCU enabled. Aug 13 00:26:46.985359 kernel: Rude variant of Tasks RCU enabled. Aug 13 00:26:46.985369 kernel: Tracing variant of Tasks RCU enabled. Aug 13 00:26:46.985378 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 00:26:46.985386 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Aug 13 00:26:46.985395 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Aug 13 00:26:46.985404 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Aug 13 00:26:46.985412 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Aug 13 00:26:46.985420 kernel: Using NULL legacy PIC Aug 13 00:26:46.985430 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 0 Aug 13 00:26:46.985438 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Aug 13 00:26:46.985447 kernel: Console: colour dummy device 80x25 Aug 13 00:26:46.985455 kernel: printk: legacy console [tty1] enabled Aug 13 00:26:46.985464 kernel: printk: legacy console [ttyS0] enabled Aug 13 00:26:46.985473 kernel: printk: legacy bootconsole [earlyser0] disabled Aug 13 00:26:46.985481 kernel: ACPI: Core revision 20240827 Aug 13 00:26:46.985491 kernel: Failed to register legacy timer interrupt Aug 13 00:26:46.985499 kernel: APIC: Switch to symmetric I/O mode setup Aug 13 00:26:46.985507 kernel: x2apic enabled Aug 13 00:26:46.985516 kernel: APIC: Switched APIC routing to: physical x2apic Aug 13 00:26:46.985524 kernel: Hyper-V: Host Build 10.0.26100.1293-1-0 Aug 13 00:26:46.985532 kernel: Hyper-V: enabling crash_kexec_post_notifiers Aug 13 00:26:46.985541 kernel: Hyper-V: Disabling IBT because of Hyper-V bug Aug 13 00:26:46.985550 kernel: Hyper-V: Using IPI hypercalls Aug 13 00:26:46.985558 kernel: APIC: send_IPI() replaced with hv_send_ipi() Aug 13 00:26:46.985568 kernel: APIC: send_IPI_mask() replaced with hv_send_ipi_mask() Aug 13 00:26:46.985577 kernel: APIC: send_IPI_mask_allbutself() replaced with hv_send_ipi_mask_allbutself() Aug 13 00:26:46.985586 kernel: APIC: send_IPI_allbutself() replaced with hv_send_ipi_allbutself() Aug 13 00:26:46.985594 kernel: APIC: send_IPI_all() replaced with hv_send_ipi_all() Aug 13 00:26:46.985838 kernel: APIC: send_IPI_self() replaced with hv_send_ipi_self() Aug 13 00:26:46.985849 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x212735223b2, max_idle_ns: 440795277976 ns Aug 13 00:26:46.985858 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 4600.00 BogoMIPS (lpj=2300000) Aug 13 00:26:46.985868 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Aug 13 00:26:46.985878 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Aug 13 00:26:46.985886 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Aug 13 00:26:46.985894 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Aug 13 00:26:46.985902 kernel: Spectre V2 : Mitigation: Retpolines Aug 13 00:26:46.985910 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Aug 13 00:26:46.985918 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Aug 13 00:26:46.985926 kernel: RETBleed: Vulnerable Aug 13 00:26:46.985935 kernel: Speculative Store Bypass: Vulnerable Aug 13 00:26:46.985943 kernel: ITS: Mitigation: Aligned branch/return thunks Aug 13 00:26:46.985952 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Aug 13 00:26:46.985960 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Aug 13 00:26:46.985970 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Aug 13 00:26:46.985978 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Aug 13 00:26:46.985987 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Aug 13 00:26:46.985995 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Aug 13 00:26:46.986003 kernel: x86/fpu: Supporting XSAVE feature 0x800: 'Control-flow User registers' Aug 13 00:26:46.986011 kernel: x86/fpu: Supporting XSAVE feature 0x20000: 'AMX Tile config' Aug 13 00:26:46.986020 kernel: x86/fpu: Supporting XSAVE feature 0x40000: 'AMX Tile data' Aug 13 00:26:46.986028 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Aug 13 00:26:46.986036 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Aug 13 00:26:46.987300 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Aug 13 00:26:46.987314 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Aug 13 00:26:46.987330 kernel: x86/fpu: xstate_offset[11]: 2432, xstate_sizes[11]: 16 Aug 13 00:26:46.987338 kernel: x86/fpu: xstate_offset[17]: 2496, xstate_sizes[17]: 64 Aug 13 00:26:46.987347 kernel: x86/fpu: xstate_offset[18]: 2560, xstate_sizes[18]: 8192 Aug 13 00:26:46.987355 kernel: x86/fpu: Enabled xstate features 0x608e7, context size is 10752 bytes, using 'compacted' format. Aug 13 00:26:46.987364 kernel: Freeing SMP alternatives memory: 32K Aug 13 00:26:46.987372 kernel: pid_max: default: 32768 minimum: 301 Aug 13 00:26:46.987380 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Aug 13 00:26:46.987388 kernel: landlock: Up and running. Aug 13 00:26:46.987397 kernel: SELinux: Initializing. Aug 13 00:26:46.987405 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Aug 13 00:26:46.987413 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Aug 13 00:26:46.987421 kernel: smpboot: CPU0: Intel INTEL(R) XEON(R) PLATINUM 8573C (family: 0x6, model: 0xcf, stepping: 0x2) Aug 13 00:26:46.987431 kernel: Performance Events: unsupported p6 CPU model 207 no PMU driver, software events only. Aug 13 00:26:46.987439 kernel: signal: max sigframe size: 11952 Aug 13 00:26:46.987447 kernel: rcu: Hierarchical SRCU implementation. Aug 13 00:26:46.987457 kernel: rcu: Max phase no-delay instances is 400. Aug 13 00:26:46.987465 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Aug 13 00:26:46.987473 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Aug 13 00:26:46.987482 kernel: smp: Bringing up secondary CPUs ... Aug 13 00:26:46.987490 kernel: smpboot: x86: Booting SMP configuration: Aug 13 00:26:46.987498 kernel: .... node #0, CPUs: #1 Aug 13 00:26:46.987508 kernel: smp: Brought up 1 node, 2 CPUs Aug 13 00:26:46.987516 kernel: smpboot: Total of 2 processors activated (9200.00 BogoMIPS) Aug 13 00:26:46.987524 kernel: Memory: 8077024K/8383228K available (14336K kernel code, 2430K rwdata, 9960K rodata, 54444K init, 2524K bss, 299988K reserved, 0K cma-reserved) Aug 13 00:26:46.987533 kernel: devtmpfs: initialized Aug 13 00:26:46.987541 kernel: x86/mm: Memory block size: 128MB Aug 13 00:26:46.987549 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x3fffb000-0x3fffefff] (16384 bytes) Aug 13 00:26:46.987557 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 00:26:46.987566 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Aug 13 00:26:46.987574 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 00:26:46.987584 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 00:26:46.987592 kernel: audit: initializing netlink subsys (disabled) Aug 13 00:26:46.987600 kernel: audit: type=2000 audit(1755044803.058:1): state=initialized audit_enabled=0 res=1 Aug 13 00:26:46.987609 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 00:26:46.987617 kernel: thermal_sys: Registered thermal governor 'user_space' Aug 13 00:26:46.987626 kernel: cpuidle: using governor menu Aug 13 00:26:46.987634 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 00:26:46.987643 kernel: dca service started, version 1.12.1 Aug 13 00:26:46.987651 kernel: e820: reserve RAM buffer [mem 0x044fe000-0x07ffffff] Aug 13 00:26:46.987661 kernel: e820: reserve RAM buffer [mem 0x3ff1f000-0x3fffffff] Aug 13 00:26:46.987669 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Aug 13 00:26:46.987677 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 00:26:46.987686 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Aug 13 00:26:46.987694 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 00:26:46.987702 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Aug 13 00:26:46.987710 kernel: ACPI: Added _OSI(Module Device) Aug 13 00:26:46.987719 kernel: ACPI: Added _OSI(Processor Device) Aug 13 00:26:46.987729 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 00:26:46.987737 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Aug 13 00:26:46.987745 kernel: ACPI: Interpreter enabled Aug 13 00:26:46.987753 kernel: ACPI: PM: (supports S0 S5) Aug 13 00:26:46.987761 kernel: ACPI: Using IOAPIC for interrupt routing Aug 13 00:26:46.987769 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Aug 13 00:26:46.987778 kernel: PCI: Ignoring E820 reservations for host bridge windows Aug 13 00:26:46.987786 kernel: ACPI: Enabled 1 GPEs in block 00 to 0F Aug 13 00:26:46.987794 kernel: iommu: Default domain type: Translated Aug 13 00:26:46.987803 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Aug 13 00:26:46.987813 kernel: efivars: Registered efivars operations Aug 13 00:26:46.987820 kernel: PCI: Using ACPI for IRQ routing Aug 13 00:26:46.987828 kernel: PCI: System does not support PCI Aug 13 00:26:46.987837 kernel: vgaarb: loaded Aug 13 00:26:46.987845 kernel: clocksource: Switched to clocksource tsc-early Aug 13 00:26:46.987854 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 00:26:46.987862 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 00:26:46.987871 kernel: pnp: PnP ACPI init Aug 13 00:26:46.987879 kernel: pnp: PnP ACPI: found 3 devices Aug 13 00:26:46.987889 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Aug 13 00:26:46.987897 kernel: NET: Registered PF_INET protocol family Aug 13 00:26:46.987906 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear) Aug 13 00:26:46.987914 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear) Aug 13 00:26:46.987922 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 00:26:46.987931 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear) Aug 13 00:26:46.987939 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Aug 13 00:26:46.987947 kernel: TCP: Hash tables configured (established 65536 bind 65536) Aug 13 00:26:46.987957 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear) Aug 13 00:26:46.987966 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear) Aug 13 00:26:46.987974 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 00:26:46.987982 kernel: NET: Registered PF_XDP protocol family Aug 13 00:26:46.987989 kernel: PCI: CLS 0 bytes, default 64 Aug 13 00:26:46.987997 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Aug 13 00:26:46.988006 kernel: software IO TLB: mapped [mem 0x000000003a9da000-0x000000003e9da000] (64MB) Aug 13 00:26:46.988014 kernel: RAPL PMU: API unit is 2^-32 Joules, 1 fixed counters, 10737418240 ms ovfl timer Aug 13 00:26:46.988022 kernel: RAPL PMU: hw unit of domain psys 2^-0 Joules Aug 13 00:26:46.988032 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x212735223b2, max_idle_ns: 440795277976 ns Aug 13 00:26:46.988041 kernel: clocksource: Switched to clocksource tsc Aug 13 00:26:46.988048 kernel: Initialise system trusted keyrings Aug 13 00:26:46.988057 kernel: workingset: timestamp_bits=39 max_order=21 bucket_order=0 Aug 13 00:26:46.988064 kernel: Key type asymmetric registered Aug 13 00:26:46.988072 kernel: Asymmetric key parser 'x509' registered Aug 13 00:26:46.988081 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Aug 13 00:26:46.988089 kernel: io scheduler mq-deadline registered Aug 13 00:26:46.988097 kernel: io scheduler kyber registered Aug 13 00:26:46.988107 kernel: io scheduler bfq registered Aug 13 00:26:46.988115 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Aug 13 00:26:46.988124 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 00:26:46.988146 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Aug 13 00:26:46.988154 kernel: 00:01: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Aug 13 00:26:46.988162 kernel: serial8250: ttyS2 at I/O 0x3e8 (irq = 4, base_baud = 115200) is a 16550A Aug 13 00:26:46.988168 kernel: i8042: PNP: No PS/2 controller found. Aug 13 00:26:46.988305 kernel: rtc_cmos 00:02: registered as rtc0 Aug 13 00:26:46.988377 kernel: rtc_cmos 00:02: setting system clock to 2025-08-13T00:26:46 UTC (1755044806) Aug 13 00:26:46.988443 kernel: rtc_cmos 00:02: alarms up to one month, 114 bytes nvram Aug 13 00:26:46.988450 kernel: intel_pstate: Intel P-state driver initializing Aug 13 00:26:46.988456 kernel: efifb: probing for efifb Aug 13 00:26:46.988462 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Aug 13 00:26:46.988474 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Aug 13 00:26:46.988483 kernel: efifb: scrolling: redraw Aug 13 00:26:46.988492 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Aug 13 00:26:46.988498 kernel: Console: switching to colour frame buffer device 128x48 Aug 13 00:26:46.988505 kernel: fb0: EFI VGA frame buffer device Aug 13 00:26:46.988510 kernel: pstore: Using crash dump compression: deflate Aug 13 00:26:46.988520 kernel: pstore: Registered efi_pstore as persistent store backend Aug 13 00:26:46.988530 kernel: NET: Registered PF_INET6 protocol family Aug 13 00:26:46.988539 kernel: Segment Routing with IPv6 Aug 13 00:26:46.988548 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 00:26:46.988555 kernel: NET: Registered PF_PACKET protocol family Aug 13 00:26:46.988560 kernel: Key type dns_resolver registered Aug 13 00:26:46.988565 kernel: IPI shorthand broadcast: enabled Aug 13 00:26:46.988573 kernel: sched_clock: Marking stable (2961159583, 93134035)->(3347270954, -292977336) Aug 13 00:26:46.988583 kernel: registered taskstats version 1 Aug 13 00:26:46.988592 kernel: Loading compiled-in X.509 certificates Aug 13 00:26:46.988600 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.40-flatcar: dee0b464d3f7f8d09744a2392f69dde258bc95c0' Aug 13 00:26:46.988607 kernel: Demotion targets for Node 0: null Aug 13 00:26:46.988612 kernel: Key type .fscrypt registered Aug 13 00:26:46.988617 kernel: Key type fscrypt-provisioning registered Aug 13 00:26:46.988624 kernel: ima: No TPM chip found, activating TPM-bypass! Aug 13 00:26:46.988632 kernel: ima: Allocated hash algorithm: sha1 Aug 13 00:26:46.988645 kernel: ima: No architecture policies found Aug 13 00:26:46.988653 kernel: clk: Disabling unused clocks Aug 13 00:26:46.988658 kernel: Warning: unable to open an initial console. Aug 13 00:26:46.988664 kernel: Freeing unused kernel image (initmem) memory: 54444K Aug 13 00:26:46.988669 kernel: Write protecting the kernel read-only data: 24576k Aug 13 00:26:46.988678 kernel: Freeing unused kernel image (rodata/data gap) memory: 280K Aug 13 00:26:46.988687 kernel: Run /init as init process Aug 13 00:26:46.988694 kernel: with arguments: Aug 13 00:26:46.988703 kernel: /init Aug 13 00:26:46.988712 kernel: with environment: Aug 13 00:26:46.988717 kernel: HOME=/ Aug 13 00:26:46.988722 kernel: TERM=linux Aug 13 00:26:46.988727 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 00:26:46.988734 systemd[1]: Successfully made /usr/ read-only. Aug 13 00:26:46.988747 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Aug 13 00:26:46.988757 systemd[1]: Detected virtualization microsoft. Aug 13 00:26:46.988766 systemd[1]: Detected architecture x86-64. Aug 13 00:26:46.988772 systemd[1]: Running in initrd. Aug 13 00:26:46.988778 systemd[1]: No hostname configured, using default hostname. Aug 13 00:26:46.988784 systemd[1]: Hostname set to . Aug 13 00:26:46.988794 systemd[1]: Initializing machine ID from random generator. Aug 13 00:26:46.988804 systemd[1]: Queued start job for default target initrd.target. Aug 13 00:26:46.988812 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 00:26:46.988817 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 00:26:46.988825 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Aug 13 00:26:46.988834 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 00:26:46.988844 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Aug 13 00:26:46.988852 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Aug 13 00:26:46.988860 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Aug 13 00:26:46.988866 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Aug 13 00:26:46.988877 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 00:26:46.988891 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 00:26:46.988900 systemd[1]: Reached target paths.target - Path Units. Aug 13 00:26:46.988906 systemd[1]: Reached target slices.target - Slice Units. Aug 13 00:26:46.988912 systemd[1]: Reached target swap.target - Swaps. Aug 13 00:26:46.988918 systemd[1]: Reached target timers.target - Timer Units. Aug 13 00:26:46.988927 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 00:26:46.988937 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 00:26:46.988945 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Aug 13 00:26:46.988952 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Aug 13 00:26:46.988959 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 00:26:46.988965 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 00:26:46.988974 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 00:26:46.988985 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 00:26:46.988993 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Aug 13 00:26:46.988999 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 00:26:46.989005 systemd[1]: Finished network-cleanup.service - Network Cleanup. Aug 13 00:26:46.989012 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Aug 13 00:26:46.989024 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 00:26:46.989033 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 00:26:46.989039 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 00:26:46.989053 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:26:46.989067 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Aug 13 00:26:46.989094 systemd-journald[205]: Collecting audit messages is disabled. Aug 13 00:26:46.989117 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 00:26:46.989131 systemd-journald[205]: Journal started Aug 13 00:26:46.989150 systemd-journald[205]: Runtime Journal (/run/log/journal/2cb8a60191eb4d4f8b746e3e73ed2bf3) is 8M, max 158.9M, 150.9M free. Aug 13 00:26:46.998926 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 00:26:47.000360 systemd-modules-load[206]: Inserted module 'overlay' Aug 13 00:26:47.000498 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 00:26:47.005862 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:26:47.011293 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 00:26:47.015947 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 00:26:47.022360 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 00:26:47.035276 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 00:26:47.038141 systemd-modules-load[206]: Inserted module 'br_netfilter' Aug 13 00:26:47.038275 kernel: Bridge firewalling registered Aug 13 00:26:47.039197 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 00:26:47.040099 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 00:26:47.050712 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 00:26:47.052398 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 00:26:47.056816 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 00:26:47.066725 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:26:47.067693 systemd-tmpfiles[221]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Aug 13 00:26:47.074141 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 00:26:47.075011 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 00:26:47.081956 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Aug 13 00:26:47.093352 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 00:26:47.104881 dracut-cmdline[244]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=215bdedb8de38f6b96ec4f9db80853e25015f60454b867e319fdcb9244320a21 Aug 13 00:26:47.137571 systemd-resolved[245]: Positive Trust Anchors: Aug 13 00:26:47.137582 systemd-resolved[245]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 00:26:47.137612 systemd-resolved[245]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 00:26:47.156918 systemd-resolved[245]: Defaulting to hostname 'linux'. Aug 13 00:26:47.159859 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 00:26:47.165051 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 00:26:47.176276 kernel: SCSI subsystem initialized Aug 13 00:26:47.184275 kernel: Loading iSCSI transport class v2.0-870. Aug 13 00:26:47.193286 kernel: iscsi: registered transport (tcp) Aug 13 00:26:47.209614 kernel: iscsi: registered transport (qla4xxx) Aug 13 00:26:47.209654 kernel: QLogic iSCSI HBA Driver Aug 13 00:26:47.222167 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Aug 13 00:26:47.234345 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Aug 13 00:26:47.236155 systemd[1]: Reached target network-pre.target - Preparation for Network. Aug 13 00:26:47.267399 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Aug 13 00:26:47.269360 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Aug 13 00:26:47.309292 kernel: raid6: avx512x4 gen() 44531 MB/s Aug 13 00:26:47.327270 kernel: raid6: avx512x2 gen() 44576 MB/s Aug 13 00:26:47.344270 kernel: raid6: avx512x1 gen() 26916 MB/s Aug 13 00:26:47.362270 kernel: raid6: avx2x4 gen() 38822 MB/s Aug 13 00:26:47.379268 kernel: raid6: avx2x2 gen() 41241 MB/s Aug 13 00:26:47.397363 kernel: raid6: avx2x1 gen() 31763 MB/s Aug 13 00:26:47.397378 kernel: raid6: using algorithm avx512x2 gen() 44576 MB/s Aug 13 00:26:47.416382 kernel: raid6: .... xor() 31547 MB/s, rmw enabled Aug 13 00:26:47.416402 kernel: raid6: using avx512x2 recovery algorithm Aug 13 00:26:47.433275 kernel: xor: automatically using best checksumming function avx Aug 13 00:26:47.545286 kernel: Btrfs loaded, zoned=no, fsverity=no Aug 13 00:26:47.549380 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Aug 13 00:26:47.552425 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 00:26:47.576565 systemd-udevd[454]: Using default interface naming scheme 'v255'. Aug 13 00:26:47.581094 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 00:26:47.587439 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Aug 13 00:26:47.609135 dracut-pre-trigger[465]: rd.md=0: removing MD RAID activation Aug 13 00:26:47.627631 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 00:26:47.631832 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 00:26:47.657836 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 00:26:47.663083 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Aug 13 00:26:47.700286 kernel: cryptd: max_cpu_qlen set to 1000 Aug 13 00:26:47.712274 kernel: AES CTR mode by8 optimization enabled Aug 13 00:26:47.734619 kernel: hv_vmbus: Vmbus version:5.3 Aug 13 00:26:47.729611 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:26:47.732497 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:26:47.736004 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:26:47.751530 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:26:47.763278 kernel: pps_core: LinuxPPS API ver. 1 registered Aug 13 00:26:47.763316 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Aug 13 00:26:47.780551 kernel: hv_vmbus: registering driver hyperv_keyboard Aug 13 00:26:47.783288 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Aug 13 00:26:47.785301 kernel: PTP clock support registered Aug 13 00:26:47.790302 kernel: hv_vmbus: registering driver hv_pci Aug 13 00:26:47.799688 kernel: hv_utils: Registering HyperV Utility Driver Aug 13 00:26:47.799725 kernel: hv_vmbus: registering driver hv_utils Aug 13 00:26:47.799739 kernel: hv_pci 7ad35d50-c05b-47ab-b3a0-56a9a845852b: PCI VMBus probing: Using version 0x10004 Aug 13 00:26:47.808280 kernel: hv_utils: Shutdown IC version 3.2 Aug 13 00:26:47.809033 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:26:47.598311 kernel: hv_utils: TimeSync IC version 4.0 Aug 13 00:26:47.602999 kernel: hv_utils: Heartbeat IC version 3.0 Aug 13 00:26:47.603014 kernel: hv_pci 7ad35d50-c05b-47ab-b3a0-56a9a845852b: PCI host bridge to bus c05b:00 Aug 13 00:26:47.603106 kernel: pci_bus c05b:00: root bus resource [mem 0xfc0000000-0xfc007ffff window] Aug 13 00:26:47.603192 kernel: pci_bus c05b:00: No busn resource found for root bus, will use [bus 00-ff] Aug 13 00:26:47.603265 kernel: pci c05b:00:00.0: [1414:00a9] type 00 class 0x010802 PCIe Endpoint Aug 13 00:26:47.603286 systemd-journald[205]: Time jumped backwards, rotating. Aug 13 00:26:47.607135 kernel: hv_vmbus: registering driver hv_netvsc Aug 13 00:26:47.607149 kernel: pci c05b:00:00.0: BAR 0 [mem 0xfc0000000-0xfc007ffff 64bit] Aug 13 00:26:47.590884 systemd-resolved[245]: Clock change detected. Flushing caches. Aug 13 00:26:47.609919 kernel: hid: raw HID events driver (C) Jiri Kosina Aug 13 00:26:47.622738 kernel: pci c05b:00:00.0: 32.000 Gb/s available PCIe bandwidth, limited by 2.5 GT/s PCIe x16 link at c05b:00:00.0 (capable of 1024.000 Gb/s with 64.0 GT/s PCIe x16 link) Aug 13 00:26:47.631648 kernel: pci_bus c05b:00: busn_res: [bus 00-ff] end is updated to 00 Aug 13 00:26:47.631846 kernel: hv_netvsc f8615163-0000-1000-2000-7ced8d403732 (unnamed net_device) (uninitialized): VF slot 1 added Aug 13 00:26:47.631977 kernel: pci c05b:00:00.0: BAR 0 [mem 0xfc0000000-0xfc007ffff 64bit]: assigned Aug 13 00:26:47.635191 kernel: hv_vmbus: registering driver hv_storvsc Aug 13 00:26:47.637687 kernel: hv_vmbus: registering driver hid_hyperv Aug 13 00:26:47.640344 kernel: scsi host0: storvsc_host_t Aug 13 00:26:47.644886 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Aug 13 00:26:47.644919 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Aug 13 00:26:47.646444 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 5 Aug 13 00:26:47.663748 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Aug 13 00:26:47.663924 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Aug 13 00:26:47.665325 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Aug 13 00:26:47.670662 kernel: nvme nvme0: pci function c05b:00:00.0 Aug 13 00:26:47.670835 kernel: nvme c05b:00:00.0: enabling device (0000 -> 0002) Aug 13 00:26:47.681309 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#108 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Aug 13 00:26:47.696311 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#235 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Aug 13 00:26:47.832348 kernel: nvme nvme0: 2/0/0 default/read/poll queues Aug 13 00:26:47.837501 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:26:48.115312 kernel: nvme nvme0: using unchecked data buffer Aug 13 00:26:48.293137 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - MSFT NVMe Accelerator v1.0 EFI-SYSTEM. Aug 13 00:26:48.361650 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - MSFT NVMe Accelerator v1.0 ROOT. Aug 13 00:26:48.370472 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - MSFT NVMe Accelerator v1.0 USR-A. Aug 13 00:26:48.370852 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - MSFT NVMe Accelerator v1.0 USR-A. Aug 13 00:26:48.377402 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Aug 13 00:26:48.430092 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - MSFT NVMe Accelerator v1.0 OEM. Aug 13 00:26:48.516649 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Aug 13 00:26:48.517810 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 00:26:48.522348 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 00:26:48.526330 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 00:26:48.530329 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Aug 13 00:26:48.545364 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Aug 13 00:26:48.653851 kernel: hv_pci 00000001-7870-47b5-b203-907d12ca697e: PCI VMBus probing: Using version 0x10004 Aug 13 00:26:48.654043 kernel: hv_pci 00000001-7870-47b5-b203-907d12ca697e: PCI host bridge to bus 7870:00 Aug 13 00:26:48.656522 kernel: pci_bus 7870:00: root bus resource [mem 0xfc2000000-0xfc4007fff window] Aug 13 00:26:48.657800 kernel: pci_bus 7870:00: No busn resource found for root bus, will use [bus 00-ff] Aug 13 00:26:48.662554 kernel: pci 7870:00:00.0: [1414:00ba] type 00 class 0x020000 PCIe Endpoint Aug 13 00:26:48.666380 kernel: pci 7870:00:00.0: BAR 0 [mem 0xfc2000000-0xfc3ffffff 64bit pref] Aug 13 00:26:48.671387 kernel: pci 7870:00:00.0: BAR 4 [mem 0xfc4000000-0xfc4007fff 64bit pref] Aug 13 00:26:48.673327 kernel: pci 7870:00:00.0: enabling Extended Tags Aug 13 00:26:48.692312 kernel: pci_bus 7870:00: busn_res: [bus 00-ff] end is updated to 00 Aug 13 00:26:48.692457 kernel: pci 7870:00:00.0: BAR 0 [mem 0xfc2000000-0xfc3ffffff 64bit pref]: assigned Aug 13 00:26:48.697387 kernel: pci 7870:00:00.0: BAR 4 [mem 0xfc4000000-0xfc4007fff 64bit pref]: assigned Aug 13 00:26:48.702186 kernel: mana 7870:00:00.0: enabling device (0000 -> 0002) Aug 13 00:26:48.713318 kernel: mana 7870:00:00.0: Microsoft Azure Network Adapter protocol version: 0.1.1 Aug 13 00:26:48.715304 kernel: hv_netvsc f8615163-0000-1000-2000-7ced8d403732 eth0: VF registering: eth1 Aug 13 00:26:48.715440 kernel: mana 7870:00:00.0 eth1: joined to eth0 Aug 13 00:26:48.720309 kernel: mana 7870:00:00.0 enP30832s1: renamed from eth1 Aug 13 00:26:49.424337 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:26:49.424395 disk-uuid[660]: The operation has completed successfully. Aug 13 00:26:49.472623 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 00:26:49.472709 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Aug 13 00:26:49.501162 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Aug 13 00:26:49.515358 sh[711]: Success Aug 13 00:26:49.545579 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 00:26:49.545628 kernel: device-mapper: uevent: version 1.0.3 Aug 13 00:26:49.546679 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Aug 13 00:26:49.555316 kernel: device-mapper: verity: sha256 using shash "sha256-ni" Aug 13 00:26:49.773870 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Aug 13 00:26:49.780382 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Aug 13 00:26:49.797196 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Aug 13 00:26:49.810144 kernel: BTRFS info: 'norecovery' is for compatibility only, recommended to use 'rescue=nologreplay' Aug 13 00:26:49.810190 kernel: BTRFS: device fsid 0c0338fb-9434-41c1-99a2-737cbe2351c4 devid 1 transid 44 /dev/mapper/usr (254:0) scanned by mount (724) Aug 13 00:26:49.813750 kernel: BTRFS info (device dm-0): first mount of filesystem 0c0338fb-9434-41c1-99a2-737cbe2351c4 Aug 13 00:26:49.813843 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Aug 13 00:26:49.814869 kernel: BTRFS info (device dm-0): using free-space-tree Aug 13 00:26:50.155531 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Aug 13 00:26:50.158128 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Aug 13 00:26:50.160027 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Aug 13 00:26:50.160759 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Aug 13 00:26:50.174065 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Aug 13 00:26:50.196757 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 (259:5) scanned by mount (747) Aug 13 00:26:50.196789 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 900bf3f4-cc50-4925-b275-d85854bb916f Aug 13 00:26:50.196805 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Aug 13 00:26:50.196815 kernel: BTRFS info (device nvme0n1p6): using free-space-tree Aug 13 00:26:50.220174 systemd[1]: Finished ignition-setup.service - Ignition (setup). Aug 13 00:26:50.225071 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 900bf3f4-cc50-4925-b275-d85854bb916f Aug 13 00:26:50.226404 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Aug 13 00:26:50.251979 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 00:26:50.254574 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 00:26:50.285153 systemd-networkd[893]: lo: Link UP Aug 13 00:26:50.285161 systemd-networkd[893]: lo: Gained carrier Aug 13 00:26:50.291362 kernel: mana 7870:00:00.0 enP30832s1: Configured vPort 0 PD 18 DB 16 Aug 13 00:26:50.286160 systemd-networkd[893]: Enumeration completed Aug 13 00:26:50.296372 kernel: mana 7870:00:00.0 enP30832s1: Configured steering vPort 0 entries 64 Aug 13 00:26:50.296567 kernel: hv_netvsc f8615163-0000-1000-2000-7ced8d403732 eth0: Data path switched to VF: enP30832s1 Aug 13 00:26:50.286365 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 00:26:50.286675 systemd-networkd[893]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 00:26:50.286678 systemd-networkd[893]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:26:50.290426 systemd[1]: Reached target network.target - Network. Aug 13 00:26:50.296975 systemd-networkd[893]: enP30832s1: Link UP Aug 13 00:26:50.297040 systemd-networkd[893]: eth0: Link UP Aug 13 00:26:50.297184 systemd-networkd[893]: eth0: Gained carrier Aug 13 00:26:50.297196 systemd-networkd[893]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 00:26:50.299632 systemd-networkd[893]: enP30832s1: Gained carrier Aug 13 00:26:50.307331 systemd-networkd[893]: eth0: DHCPv4 address 10.200.8.20/24, gateway 10.200.8.1 acquired from 168.63.129.16 Aug 13 00:26:51.120752 ignition[842]: Ignition 2.21.0 Aug 13 00:26:51.120763 ignition[842]: Stage: fetch-offline Aug 13 00:26:51.122667 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 00:26:51.120850 ignition[842]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:26:51.128683 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Aug 13 00:26:51.120857 ignition[842]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 00:26:51.120941 ignition[842]: parsed url from cmdline: "" Aug 13 00:26:51.120943 ignition[842]: no config URL provided Aug 13 00:26:51.120947 ignition[842]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 00:26:51.120952 ignition[842]: no config at "/usr/lib/ignition/user.ign" Aug 13 00:26:51.120956 ignition[842]: failed to fetch config: resource requires networking Aug 13 00:26:51.121588 ignition[842]: Ignition finished successfully Aug 13 00:26:51.150505 ignition[904]: Ignition 2.21.0 Aug 13 00:26:51.150513 ignition[904]: Stage: fetch Aug 13 00:26:51.150676 ignition[904]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:26:51.150683 ignition[904]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 00:26:51.150745 ignition[904]: parsed url from cmdline: "" Aug 13 00:26:51.150748 ignition[904]: no config URL provided Aug 13 00:26:51.150751 ignition[904]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 00:26:51.150756 ignition[904]: no config at "/usr/lib/ignition/user.ign" Aug 13 00:26:51.150792 ignition[904]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Aug 13 00:26:51.225537 ignition[904]: GET result: OK Aug 13 00:26:51.225663 ignition[904]: config has been read from IMDS userdata Aug 13 00:26:51.225702 ignition[904]: parsing config with SHA512: e3118d9ee551cb742c879a8f3689bdf1faf9fabd66abdfa88dad74745dc446c12a04f209c0ba1834f1f00292f1d9cfd08e522d21c2c53099ea5b5909be289c30 Aug 13 00:26:51.229136 unknown[904]: fetched base config from "system" Aug 13 00:26:51.229174 unknown[904]: fetched base config from "system" Aug 13 00:26:51.229481 ignition[904]: fetch: fetch complete Aug 13 00:26:51.229179 unknown[904]: fetched user config from "azure" Aug 13 00:26:51.229485 ignition[904]: fetch: fetch passed Aug 13 00:26:51.231379 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Aug 13 00:26:51.229520 ignition[904]: Ignition finished successfully Aug 13 00:26:51.235240 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Aug 13 00:26:51.268321 ignition[911]: Ignition 2.21.0 Aug 13 00:26:51.268330 ignition[911]: Stage: kargs Aug 13 00:26:51.270803 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Aug 13 00:26:51.268557 ignition[911]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:26:51.274091 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Aug 13 00:26:51.268572 ignition[911]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 00:26:51.269372 ignition[911]: kargs: kargs passed Aug 13 00:26:51.269411 ignition[911]: Ignition finished successfully Aug 13 00:26:51.310009 ignition[918]: Ignition 2.21.0 Aug 13 00:26:51.310019 ignition[918]: Stage: disks Aug 13 00:26:51.310198 ignition[918]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:26:51.311803 systemd[1]: Finished ignition-disks.service - Ignition (disks). Aug 13 00:26:51.310205 ignition[918]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 00:26:51.314090 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Aug 13 00:26:51.310953 ignition[918]: disks: disks passed Aug 13 00:26:51.310982 ignition[918]: Ignition finished successfully Aug 13 00:26:51.323364 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Aug 13 00:26:51.326084 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 00:26:51.329329 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 00:26:51.333331 systemd[1]: Reached target basic.target - Basic System. Aug 13 00:26:51.336666 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Aug 13 00:26:51.424504 systemd-fsck[926]: ROOT: clean, 15/7326000 files, 477845/7359488 blocks Aug 13 00:26:51.428182 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Aug 13 00:26:51.432364 systemd[1]: Mounting sysroot.mount - /sysroot... Aug 13 00:26:51.610415 systemd-networkd[893]: eth0: Gained IPv6LL Aug 13 00:26:51.710310 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 069caac6-7833-4acd-8940-01a7ff7d1281 r/w with ordered data mode. Quota mode: none. Aug 13 00:26:51.710283 systemd[1]: Mounted sysroot.mount - /sysroot. Aug 13 00:26:51.713785 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Aug 13 00:26:51.730790 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 00:26:51.732951 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Aug 13 00:26:51.746403 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Aug 13 00:26:51.750090 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 00:26:51.750124 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 00:26:51.763217 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 (259:5) scanned by mount (935) Aug 13 00:26:51.755062 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Aug 13 00:26:51.768804 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 900bf3f4-cc50-4925-b275-d85854bb916f Aug 13 00:26:51.768824 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Aug 13 00:26:51.768839 kernel: BTRFS info (device nvme0n1p6): using free-space-tree Aug 13 00:26:51.763553 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Aug 13 00:26:51.773927 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 00:26:52.407126 initrd-setup-root[960]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 00:26:52.428702 initrd-setup-root[967]: cut: /sysroot/etc/group: No such file or directory Aug 13 00:26:52.446748 initrd-setup-root[974]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 00:26:52.479814 initrd-setup-root[981]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 00:26:52.537753 coreos-metadata[937]: Aug 13 00:26:52.537 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Aug 13 00:26:52.540912 coreos-metadata[937]: Aug 13 00:26:52.540 INFO Fetch successful Aug 13 00:26:52.543341 coreos-metadata[937]: Aug 13 00:26:52.540 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Aug 13 00:26:52.550814 coreos-metadata[937]: Aug 13 00:26:52.550 INFO Fetch successful Aug 13 00:26:52.565096 coreos-metadata[937]: Aug 13 00:26:52.565 INFO wrote hostname ci-4372.1.0-a-4f0cf7bbd9 to /sysroot/etc/hostname Aug 13 00:26:52.567092 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 00:26:53.812551 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Aug 13 00:26:53.817251 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Aug 13 00:26:53.820777 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Aug 13 00:26:53.833432 systemd[1]: sysroot-oem.mount: Deactivated successfully. Aug 13 00:26:53.835849 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 900bf3f4-cc50-4925-b275-d85854bb916f Aug 13 00:26:53.857134 ignition[1054]: INFO : Ignition 2.21.0 Aug 13 00:26:53.857134 ignition[1054]: INFO : Stage: mount Aug 13 00:26:53.860258 ignition[1054]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 00:26:53.860258 ignition[1054]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 00:26:53.860550 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Aug 13 00:26:53.869714 ignition[1054]: INFO : mount: mount passed Aug 13 00:26:53.869714 ignition[1054]: INFO : Ignition finished successfully Aug 13 00:26:53.864542 systemd[1]: Finished ignition-mount.service - Ignition (mount). Aug 13 00:26:53.870376 systemd[1]: Starting ignition-files.service - Ignition (files)... Aug 13 00:26:53.888846 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 00:26:53.910306 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 (259:5) scanned by mount (1066) Aug 13 00:26:53.910337 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 900bf3f4-cc50-4925-b275-d85854bb916f Aug 13 00:26:53.912770 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Aug 13 00:26:53.912789 kernel: BTRFS info (device nvme0n1p6): using free-space-tree Aug 13 00:26:53.917446 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 00:26:53.939692 ignition[1083]: INFO : Ignition 2.21.0 Aug 13 00:26:53.939692 ignition[1083]: INFO : Stage: files Aug 13 00:26:53.942500 ignition[1083]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 00:26:53.942500 ignition[1083]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 00:26:53.942500 ignition[1083]: DEBUG : files: compiled without relabeling support, skipping Aug 13 00:26:53.947069 ignition[1083]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 00:26:53.947069 ignition[1083]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 00:26:53.961957 ignition[1083]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 00:26:53.965379 ignition[1083]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 00:26:53.965379 ignition[1083]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 00:26:53.963950 unknown[1083]: wrote ssh authorized keys file for user: core Aug 13 00:26:53.980099 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Aug 13 00:26:53.984346 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Aug 13 00:27:23.991496 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET error: Get "https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz": dial tcp 13.107.246.52:443: i/o timeout Aug 13 00:27:24.191868 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #2 Aug 13 00:27:28.359540 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Aug 13 00:27:28.401877 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Aug 13 00:27:28.404640 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Aug 13 00:27:28.408343 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 00:27:28.408343 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 00:27:28.408343 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 00:27:28.408343 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 00:27:28.408343 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 00:27:28.408343 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 00:27:28.408343 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 00:27:28.427490 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 00:27:28.427490 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 00:27:28.427490 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 00:27:28.427490 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 00:27:28.427490 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 00:27:28.427490 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Aug 13 00:27:28.951002 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Aug 13 00:27:29.560632 ignition[1083]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 00:27:29.560632 ignition[1083]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Aug 13 00:27:29.596929 ignition[1083]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 00:27:29.605152 ignition[1083]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 00:27:29.605152 ignition[1083]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Aug 13 00:27:29.605152 ignition[1083]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Aug 13 00:27:29.616973 ignition[1083]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 00:27:29.616973 ignition[1083]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 00:27:29.616973 ignition[1083]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 00:27:29.616973 ignition[1083]: INFO : files: files passed Aug 13 00:27:29.616973 ignition[1083]: INFO : Ignition finished successfully Aug 13 00:27:29.611281 systemd[1]: Finished ignition-files.service - Ignition (files). Aug 13 00:27:29.615227 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Aug 13 00:27:29.628790 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Aug 13 00:27:29.635161 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 00:27:29.636732 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Aug 13 00:27:29.650427 initrd-setup-root-after-ignition[1113]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:27:29.650427 initrd-setup-root-after-ignition[1113]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:27:29.648604 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 00:27:29.658693 initrd-setup-root-after-ignition[1117]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:27:29.652636 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Aug 13 00:27:29.666829 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Aug 13 00:27:29.691556 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 00:27:29.691635 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Aug 13 00:27:29.696624 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Aug 13 00:27:29.700352 systemd[1]: Reached target initrd.target - Initrd Default Target. Aug 13 00:27:29.704392 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Aug 13 00:27:29.707841 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Aug 13 00:27:29.720383 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 00:27:29.723413 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Aug 13 00:27:29.750434 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Aug 13 00:27:29.750578 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 00:27:29.750795 systemd[1]: Stopped target timers.target - Timer Units. Aug 13 00:27:29.756441 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 00:27:29.756584 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 00:27:29.763380 systemd[1]: Stopped target initrd.target - Initrd Default Target. Aug 13 00:27:29.765871 systemd[1]: Stopped target basic.target - Basic System. Aug 13 00:27:29.767605 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Aug 13 00:27:29.768185 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 00:27:29.768805 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Aug 13 00:27:29.769335 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Aug 13 00:27:29.769905 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Aug 13 00:27:29.770186 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 00:27:29.770518 systemd[1]: Stopped target sysinit.target - System Initialization. Aug 13 00:27:29.784613 systemd[1]: Stopped target local-fs.target - Local File Systems. Aug 13 00:27:29.788437 systemd[1]: Stopped target swap.target - Swaps. Aug 13 00:27:29.790437 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 00:27:29.790558 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Aug 13 00:27:29.794677 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Aug 13 00:27:29.798456 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 00:27:29.803388 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Aug 13 00:27:29.804461 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 00:27:29.819668 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 00:27:29.819818 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Aug 13 00:27:29.825481 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 00:27:29.825637 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 00:27:29.829098 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 00:27:29.829217 systemd[1]: Stopped ignition-files.service - Ignition (files). Aug 13 00:27:29.836448 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Aug 13 00:27:29.836598 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 00:27:29.842462 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Aug 13 00:27:29.852574 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Aug 13 00:27:29.857390 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 00:27:29.859523 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 00:27:29.863259 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 00:27:29.863382 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 00:27:29.874865 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 00:27:29.874946 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Aug 13 00:27:29.884358 ignition[1137]: INFO : Ignition 2.21.0 Aug 13 00:27:29.884358 ignition[1137]: INFO : Stage: umount Aug 13 00:27:29.884358 ignition[1137]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 00:27:29.884358 ignition[1137]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 00:27:29.884358 ignition[1137]: INFO : umount: umount passed Aug 13 00:27:29.906547 ignition[1137]: INFO : Ignition finished successfully Aug 13 00:27:29.886760 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 00:27:29.887136 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 00:27:29.887192 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Aug 13 00:27:29.892251 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 00:27:29.892341 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Aug 13 00:27:29.896413 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 00:27:29.896460 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Aug 13 00:27:29.896889 systemd[1]: ignition-fetch.service: Deactivated successfully. Aug 13 00:27:29.896914 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Aug 13 00:27:29.897415 systemd[1]: Stopped target network.target - Network. Aug 13 00:27:29.897438 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 00:27:29.897462 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 00:27:29.897892 systemd[1]: Stopped target paths.target - Path Units. Aug 13 00:27:29.905385 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 00:27:29.905457 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 00:27:29.912441 systemd[1]: Stopped target slices.target - Slice Units. Aug 13 00:27:29.913667 systemd[1]: Stopped target sockets.target - Socket Units. Aug 13 00:27:29.913735 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 00:27:29.913768 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 00:27:29.913947 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 00:27:29.913970 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 00:27:29.914187 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 00:27:29.914227 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Aug 13 00:27:29.914401 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Aug 13 00:27:29.914427 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Aug 13 00:27:29.915095 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Aug 13 00:27:29.915569 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Aug 13 00:27:29.916053 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 00:27:29.916120 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Aug 13 00:27:29.916457 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 00:27:29.916515 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Aug 13 00:27:29.929644 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 00:27:29.929732 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Aug 13 00:27:29.938798 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Aug 13 00:27:29.938926 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 00:27:29.938993 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Aug 13 00:27:29.941998 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Aug 13 00:27:29.942380 systemd[1]: Stopped target network-pre.target - Preparation for Network. Aug 13 00:27:29.943840 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 00:27:29.943877 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Aug 13 00:27:29.967910 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Aug 13 00:27:29.970814 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 00:27:29.970919 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 00:27:29.972739 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 00:27:29.972779 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 00:27:29.976589 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 00:27:29.976635 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Aug 13 00:27:29.986788 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Aug 13 00:27:29.986828 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 00:27:29.991024 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 00:27:30.014546 kernel: hv_netvsc f8615163-0000-1000-2000-7ced8d403732 eth0: Data path switched from VF: enP30832s1 Aug 13 00:27:29.997091 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Aug 13 00:27:29.997142 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Aug 13 00:27:30.000967 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 00:27:30.001093 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 00:27:30.003064 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 00:27:30.003119 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Aug 13 00:27:30.003723 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 00:27:30.003745 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 00:27:30.004117 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 00:27:30.004148 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Aug 13 00:27:30.004372 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 00:27:30.004396 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Aug 13 00:27:30.004584 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 00:27:30.004608 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:27:30.007490 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Aug 13 00:27:30.049356 kernel: mana 7870:00:00.0 enP30832s1: Configured steering vPort 0 entries 64 Aug 13 00:27:30.043386 systemd[1]: systemd-network-generator.service: Deactivated successfully. Aug 13 00:27:30.043444 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Aug 13 00:27:30.046696 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Aug 13 00:27:30.046734 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 00:27:30.050865 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Aug 13 00:27:30.050912 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 00:27:30.059036 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 00:27:30.059332 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 00:27:30.064168 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:27:30.064205 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:27:30.072049 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Aug 13 00:27:30.072086 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. Aug 13 00:27:30.072108 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Aug 13 00:27:30.072131 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Aug 13 00:27:30.072391 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 00:27:30.073859 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Aug 13 00:27:30.079521 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 00:27:30.079593 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Aug 13 00:27:30.084829 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Aug 13 00:27:30.090877 systemd[1]: Starting initrd-switch-root.service - Switch Root... Aug 13 00:27:30.121133 systemd[1]: Switching root. Aug 13 00:27:30.186713 systemd-journald[205]: Journal stopped Aug 13 00:27:34.156997 systemd-journald[205]: Received SIGTERM from PID 1 (systemd). Aug 13 00:27:34.157015 kernel: SELinux: policy capability network_peer_controls=1 Aug 13 00:27:34.157023 kernel: SELinux: policy capability open_perms=1 Aug 13 00:27:34.157029 kernel: SELinux: policy capability extended_socket_class=1 Aug 13 00:27:34.157033 kernel: SELinux: policy capability always_check_network=0 Aug 13 00:27:34.157048 kernel: SELinux: policy capability cgroup_seclabel=1 Aug 13 00:27:34.157058 kernel: SELinux: policy capability nnp_nosuid_transition=1 Aug 13 00:27:34.157067 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Aug 13 00:27:34.157075 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Aug 13 00:27:34.157083 kernel: SELinux: policy capability userspace_initial_context=0 Aug 13 00:27:34.157091 kernel: audit: type=1403 audit(1755044851.658:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Aug 13 00:27:34.157100 systemd[1]: Successfully loaded SELinux policy in 123.168ms. Aug 13 00:27:34.157110 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 6.564ms. Aug 13 00:27:34.157134 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Aug 13 00:27:34.157143 systemd[1]: Detected virtualization microsoft. Aug 13 00:27:34.157152 systemd[1]: Detected architecture x86-64. Aug 13 00:27:34.157161 systemd[1]: Detected first boot. Aug 13 00:27:34.157170 systemd[1]: Hostname set to . Aug 13 00:27:34.157180 systemd[1]: Initializing machine ID from random generator. Aug 13 00:27:34.157189 zram_generator::config[1180]: No configuration found. Aug 13 00:27:34.157199 kernel: Guest personality initialized and is inactive Aug 13 00:27:34.157208 kernel: VMCI host device registered (name=vmci, major=10, minor=124) Aug 13 00:27:34.157216 kernel: Initialized host personality Aug 13 00:27:34.157225 kernel: NET: Registered PF_VSOCK protocol family Aug 13 00:27:34.157234 systemd[1]: Populated /etc with preset unit settings. Aug 13 00:27:34.157245 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Aug 13 00:27:34.157255 systemd[1]: initrd-switch-root.service: Deactivated successfully. Aug 13 00:27:34.157266 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Aug 13 00:27:34.157275 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Aug 13 00:27:34.157284 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Aug 13 00:27:34.157325 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Aug 13 00:27:34.157336 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Aug 13 00:27:34.157347 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Aug 13 00:27:34.157356 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Aug 13 00:27:34.157365 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Aug 13 00:27:34.157374 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Aug 13 00:27:34.157383 systemd[1]: Created slice user.slice - User and Session Slice. Aug 13 00:27:34.157392 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 00:27:34.157401 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 00:27:34.157411 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Aug 13 00:27:34.157423 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Aug 13 00:27:34.157434 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Aug 13 00:27:34.157445 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 00:27:34.157455 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Aug 13 00:27:34.157464 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 00:27:34.157474 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 00:27:34.157484 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Aug 13 00:27:34.157494 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Aug 13 00:27:34.157506 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Aug 13 00:27:34.157517 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Aug 13 00:27:34.157527 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 00:27:34.157538 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 00:27:34.157548 systemd[1]: Reached target slices.target - Slice Units. Aug 13 00:27:34.157557 systemd[1]: Reached target swap.target - Swaps. Aug 13 00:27:34.157566 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Aug 13 00:27:34.157576 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Aug 13 00:27:34.157588 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Aug 13 00:27:34.157598 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 00:27:34.157607 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 00:27:34.157616 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 00:27:34.157625 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Aug 13 00:27:34.157636 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Aug 13 00:27:34.157647 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Aug 13 00:27:34.157656 systemd[1]: Mounting media.mount - External Media Directory... Aug 13 00:27:34.157665 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 00:27:34.157674 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Aug 13 00:27:34.157684 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Aug 13 00:27:34.157694 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Aug 13 00:27:34.157704 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Aug 13 00:27:34.157716 systemd[1]: Reached target machines.target - Containers. Aug 13 00:27:34.157725 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Aug 13 00:27:34.157734 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 00:27:34.157744 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 00:27:34.157754 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Aug 13 00:27:34.157763 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 00:27:34.157773 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 00:27:34.157782 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 00:27:34.157792 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Aug 13 00:27:34.157803 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 00:27:34.157813 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Aug 13 00:27:34.157822 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Aug 13 00:27:34.157832 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Aug 13 00:27:34.157842 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Aug 13 00:27:34.157851 systemd[1]: Stopped systemd-fsck-usr.service. Aug 13 00:27:34.157863 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Aug 13 00:27:34.157874 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 00:27:34.162373 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 00:27:34.162396 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Aug 13 00:27:34.162409 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Aug 13 00:27:34.162420 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Aug 13 00:27:34.162431 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 00:27:34.162441 kernel: loop: module loaded Aug 13 00:27:34.162451 systemd[1]: verity-setup.service: Deactivated successfully. Aug 13 00:27:34.162462 systemd[1]: Stopped verity-setup.service. Aug 13 00:27:34.162477 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 00:27:34.162489 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Aug 13 00:27:34.162500 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Aug 13 00:27:34.162510 kernel: fuse: init (API version 7.41) Aug 13 00:27:34.162519 systemd[1]: Mounted media.mount - External Media Directory. Aug 13 00:27:34.162529 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Aug 13 00:27:34.162564 systemd-journald[1264]: Collecting audit messages is disabled. Aug 13 00:27:34.162591 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Aug 13 00:27:34.162602 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Aug 13 00:27:34.163085 systemd-journald[1264]: Journal started Aug 13 00:27:34.163113 systemd-journald[1264]: Runtime Journal (/run/log/journal/dd6855cd3ae040efa04c2538338d7a7d) is 8M, max 158.9M, 150.9M free. Aug 13 00:27:33.780168 systemd[1]: Queued start job for default target multi-user.target. Aug 13 00:27:33.787825 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Aug 13 00:27:34.165321 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 00:27:33.788144 systemd[1]: systemd-journald.service: Deactivated successfully. Aug 13 00:27:34.168746 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 00:27:34.172721 systemd[1]: modprobe@configfs.service: Deactivated successfully. Aug 13 00:27:34.172937 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Aug 13 00:27:34.175468 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 00:27:34.175698 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 00:27:34.178539 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 00:27:34.178679 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 00:27:34.180480 systemd[1]: modprobe@fuse.service: Deactivated successfully. Aug 13 00:27:34.180617 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Aug 13 00:27:34.183621 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Aug 13 00:27:34.194675 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 00:27:34.194826 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 00:27:34.198592 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 00:27:34.200887 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Aug 13 00:27:34.203967 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Aug 13 00:27:34.206855 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Aug 13 00:27:34.215993 systemd[1]: Reached target network-pre.target - Preparation for Network. Aug 13 00:27:34.221972 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Aug 13 00:27:34.226053 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Aug 13 00:27:34.228479 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Aug 13 00:27:34.228511 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 00:27:34.232219 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Aug 13 00:27:34.237264 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Aug 13 00:27:34.240475 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 00:27:34.241469 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Aug 13 00:27:34.247922 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Aug 13 00:27:34.252400 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 00:27:34.253911 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Aug 13 00:27:34.257408 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 00:27:34.259175 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 00:27:34.263784 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Aug 13 00:27:34.268472 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 00:27:34.272797 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 00:27:34.277070 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Aug 13 00:27:34.280037 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Aug 13 00:27:34.292567 systemd-journald[1264]: Time spent on flushing to /var/log/journal/dd6855cd3ae040efa04c2538338d7a7d is 42.442ms for 983 entries. Aug 13 00:27:34.292567 systemd-journald[1264]: System Journal (/var/log/journal/dd6855cd3ae040efa04c2538338d7a7d) is 11.8M, max 2.6G, 2.6G free. Aug 13 00:27:34.449599 systemd-journald[1264]: Received client request to flush runtime journal. Aug 13 00:27:34.449648 kernel: ACPI: bus type drm_connector registered Aug 13 00:27:34.449670 systemd-journald[1264]: /var/log/journal/dd6855cd3ae040efa04c2538338d7a7d/system.journal: Realtime clock jumped backwards relative to last journal entry, rotating. Aug 13 00:27:34.449701 systemd-journald[1264]: Rotating system journal. Aug 13 00:27:34.449725 kernel: loop0: detected capacity change from 0 to 113872 Aug 13 00:27:34.299803 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 00:27:34.299978 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 00:27:34.307145 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Aug 13 00:27:34.310630 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Aug 13 00:27:34.315513 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Aug 13 00:27:34.358103 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 00:27:34.438841 systemd-tmpfiles[1322]: ACLs are not supported, ignoring. Aug 13 00:27:34.438852 systemd-tmpfiles[1322]: ACLs are not supported, ignoring. Aug 13 00:27:34.441851 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 00:27:34.446434 systemd[1]: Starting systemd-sysusers.service - Create System Users... Aug 13 00:27:34.451281 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Aug 13 00:27:34.463528 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Aug 13 00:27:34.550314 systemd[1]: Finished systemd-sysusers.service - Create System Users. Aug 13 00:27:34.552571 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 00:27:34.569769 systemd-tmpfiles[1343]: ACLs are not supported, ignoring. Aug 13 00:27:34.569901 systemd-tmpfiles[1343]: ACLs are not supported, ignoring. Aug 13 00:27:34.572024 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 00:27:34.714314 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Aug 13 00:27:34.743314 kernel: loop1: detected capacity change from 0 to 146240 Aug 13 00:27:34.789307 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Aug 13 00:27:35.138321 kernel: loop2: detected capacity change from 0 to 229808 Aug 13 00:27:35.168942 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Aug 13 00:27:35.171695 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 00:27:35.191307 kernel: loop3: detected capacity change from 0 to 28496 Aug 13 00:27:35.196806 systemd-udevd[1351]: Using default interface naming scheme 'v255'. Aug 13 00:27:35.286304 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 00:27:35.293470 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 00:27:35.358702 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Aug 13 00:27:35.368828 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Aug 13 00:27:35.419312 kernel: mousedev: PS/2 mouse device common for all mice Aug 13 00:27:35.457681 kernel: hv_vmbus: registering driver hyperv_fb Aug 13 00:27:35.462650 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Aug 13 00:27:35.467311 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Aug 13 00:27:35.469486 kernel: Console: switching to colour dummy device 80x25 Aug 13 00:27:35.475698 kernel: Console: switching to colour frame buffer device 128x48 Aug 13 00:27:35.479333 kernel: hv_vmbus: registering driver hv_balloon Aug 13 00:27:35.486780 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#207 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Aug 13 00:27:35.507011 systemd[1]: Started systemd-userdbd.service - User Database Manager. Aug 13 00:27:35.511315 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Aug 13 00:27:35.629273 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:27:35.647093 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:27:35.647459 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:27:35.654535 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:27:35.661233 systemd-networkd[1364]: lo: Link UP Aug 13 00:27:35.662331 systemd-networkd[1364]: lo: Gained carrier Aug 13 00:27:35.663750 systemd-networkd[1364]: Enumeration completed Aug 13 00:27:35.664396 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 00:27:35.667046 systemd-networkd[1364]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 00:27:35.667050 systemd-networkd[1364]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:27:35.669305 kernel: mana 7870:00:00.0 enP30832s1: Configured vPort 0 PD 18 DB 16 Aug 13 00:27:35.668328 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Aug 13 00:27:35.675825 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Aug 13 00:27:35.680317 kernel: mana 7870:00:00.0 enP30832s1: Configured steering vPort 0 entries 64 Aug 13 00:27:35.686142 kernel: hv_netvsc f8615163-0000-1000-2000-7ced8d403732 eth0: Data path switched to VF: enP30832s1 Aug 13 00:27:35.685503 systemd-networkd[1364]: enP30832s1: Link UP Aug 13 00:27:35.685570 systemd-networkd[1364]: eth0: Link UP Aug 13 00:27:35.685572 systemd-networkd[1364]: eth0: Gained carrier Aug 13 00:27:35.685587 systemd-networkd[1364]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 00:27:35.689477 systemd-networkd[1364]: enP30832s1: Gained carrier Aug 13 00:27:35.695393 systemd-networkd[1364]: eth0: DHCPv4 address 10.200.8.20/24, gateway 10.200.8.1 acquired from 168.63.129.16 Aug 13 00:27:35.705278 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:27:35.705515 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:27:35.710188 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Aug 13 00:27:35.713470 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:27:35.770446 kernel: loop4: detected capacity change from 0 to 113872 Aug 13 00:27:35.771050 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Aug 13 00:27:35.776773 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - MSFT NVMe Accelerator v1.0 OEM. Aug 13 00:27:35.783773 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Aug 13 00:27:35.799349 kernel: loop5: detected capacity change from 0 to 146240 Aug 13 00:27:35.812454 kernel: loop6: detected capacity change from 0 to 229808 Aug 13 00:27:35.826307 kernel: loop7: detected capacity change from 0 to 28496 Aug 13 00:27:35.837176 (sd-merge)[1445]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Aug 13 00:27:35.838552 (sd-merge)[1445]: Merged extensions into '/usr'. Aug 13 00:27:35.843408 systemd[1]: Reload requested from client PID 1321 ('systemd-sysext') (unit systemd-sysext.service)... Aug 13 00:27:35.843424 systemd[1]: Reloading... Aug 13 00:27:35.851320 kernel: kvm_intel: Using Hyper-V Enlightened VMCS Aug 13 00:27:35.907312 zram_generator::config[1477]: No configuration found. Aug 13 00:27:35.993734 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:27:36.085410 systemd[1]: Reloading finished in 241 ms. Aug 13 00:27:36.100068 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Aug 13 00:27:36.103616 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:27:36.106560 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Aug 13 00:27:36.115110 systemd[1]: Starting ensure-sysext.service... Aug 13 00:27:36.118321 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 00:27:36.137617 systemd[1]: Reload requested from client PID 1542 ('systemctl') (unit ensure-sysext.service)... Aug 13 00:27:36.137635 systemd[1]: Reloading... Aug 13 00:27:36.162018 systemd-tmpfiles[1543]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Aug 13 00:27:36.162046 systemd-tmpfiles[1543]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Aug 13 00:27:36.162345 systemd-tmpfiles[1543]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Aug 13 00:27:36.162567 systemd-tmpfiles[1543]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Aug 13 00:27:36.163526 systemd-tmpfiles[1543]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Aug 13 00:27:36.163759 systemd-tmpfiles[1543]: ACLs are not supported, ignoring. Aug 13 00:27:36.163812 systemd-tmpfiles[1543]: ACLs are not supported, ignoring. Aug 13 00:27:36.184365 systemd-tmpfiles[1543]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 00:27:36.184378 systemd-tmpfiles[1543]: Skipping /boot Aug 13 00:27:36.193331 zram_generator::config[1570]: No configuration found. Aug 13 00:27:36.194853 systemd-tmpfiles[1543]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 00:27:36.194868 systemd-tmpfiles[1543]: Skipping /boot Aug 13 00:27:36.282634 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:27:36.364600 systemd[1]: Reloading finished in 226 ms. Aug 13 00:27:36.386082 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 00:27:36.394009 systemd[1]: Starting audit-rules.service - Load Audit Rules... Aug 13 00:27:36.404105 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Aug 13 00:27:36.408583 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Aug 13 00:27:36.414491 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 00:27:36.418019 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Aug 13 00:27:36.424120 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 00:27:36.424266 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 00:27:36.427687 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 00:27:36.433392 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 00:27:36.437482 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 00:27:36.439899 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 00:27:36.440008 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Aug 13 00:27:36.440102 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 00:27:36.444840 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 00:27:36.445031 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 00:27:36.445216 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 00:27:36.446385 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Aug 13 00:27:36.446559 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 00:27:36.450142 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Aug 13 00:27:36.453418 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 00:27:36.453566 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 00:27:36.457778 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 00:27:36.457980 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 00:27:36.461072 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 00:27:36.461212 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 00:27:36.473214 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 00:27:36.473483 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 00:27:36.476489 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 00:27:36.479811 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 00:27:36.484518 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 00:27:36.488602 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 00:27:36.492461 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 00:27:36.493409 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Aug 13 00:27:36.493575 systemd[1]: Reached target time-set.target - System Time Set. Aug 13 00:27:36.496295 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 00:27:36.502071 systemd[1]: Finished ensure-sysext.service. Aug 13 00:27:36.508232 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 00:27:36.508413 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 00:27:36.512704 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 00:27:36.513046 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 00:27:36.516974 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 00:27:36.517166 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 00:27:36.521622 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 00:27:36.521780 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 00:27:36.525579 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Aug 13 00:27:36.530086 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 00:27:36.530134 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 00:27:36.534684 systemd-resolved[1638]: Positive Trust Anchors: Aug 13 00:27:36.534694 systemd-resolved[1638]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 00:27:36.534719 systemd-resolved[1638]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 00:27:36.551325 systemd-resolved[1638]: Using system hostname 'ci-4372.1.0-a-4f0cf7bbd9'. Aug 13 00:27:36.552741 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 00:27:36.555386 systemd[1]: Reached target network.target - Network. Aug 13 00:27:36.558330 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 00:27:36.564579 augenrules[1676]: No rules Aug 13 00:27:36.565336 systemd[1]: audit-rules.service: Deactivated successfully. Aug 13 00:27:36.565495 systemd[1]: Finished audit-rules.service - Load Audit Rules. Aug 13 00:27:36.766103 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Aug 13 00:27:36.769512 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 00:27:37.562427 systemd-networkd[1364]: eth0: Gained IPv6LL Aug 13 00:27:37.564414 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Aug 13 00:27:37.568588 systemd[1]: Reached target network-online.target - Network is Online. Aug 13 00:27:39.350123 ldconfig[1316]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Aug 13 00:27:39.358879 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Aug 13 00:27:39.361780 systemd[1]: Starting systemd-update-done.service - Update is Completed... Aug 13 00:27:39.381579 systemd[1]: Finished systemd-update-done.service - Update is Completed. Aug 13 00:27:39.384522 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 00:27:39.387451 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Aug 13 00:27:39.390362 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Aug 13 00:27:39.391863 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Aug 13 00:27:39.393499 systemd[1]: Started logrotate.timer - Daily rotation of log files. Aug 13 00:27:39.396391 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Aug 13 00:27:39.399347 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Aug 13 00:27:39.402343 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Aug 13 00:27:39.402376 systemd[1]: Reached target paths.target - Path Units. Aug 13 00:27:39.405340 systemd[1]: Reached target timers.target - Timer Units. Aug 13 00:27:39.409461 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Aug 13 00:27:39.413243 systemd[1]: Starting docker.socket - Docker Socket for the API... Aug 13 00:27:39.418275 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Aug 13 00:27:39.421478 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Aug 13 00:27:39.422830 systemd[1]: Reached target ssh-access.target - SSH Access Available. Aug 13 00:27:39.434745 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Aug 13 00:27:39.436581 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Aug 13 00:27:39.439089 systemd[1]: Listening on docker.socket - Docker Socket for the API. Aug 13 00:27:39.444237 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 00:27:39.445608 systemd[1]: Reached target basic.target - Basic System. Aug 13 00:27:39.449369 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Aug 13 00:27:39.449397 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Aug 13 00:27:39.451131 systemd[1]: Starting chronyd.service - NTP client/server... Aug 13 00:27:39.454993 systemd[1]: Starting containerd.service - containerd container runtime... Aug 13 00:27:39.458954 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Aug 13 00:27:39.463413 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Aug 13 00:27:39.467426 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Aug 13 00:27:39.474797 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Aug 13 00:27:39.478347 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Aug 13 00:27:39.480448 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Aug 13 00:27:39.483111 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Aug 13 00:27:39.486086 jq[1694]: false Aug 13 00:27:39.485542 systemd[1]: hv_fcopy_uio_daemon.service - Hyper-V FCOPY UIO daemon was skipped because of an unmet condition check (ConditionPathExists=/sys/bus/vmbus/devices/eb765408-105f-49b6-b4aa-c123b64d17d4/uio). Aug 13 00:27:39.486585 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. Aug 13 00:27:39.489433 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). Aug 13 00:27:39.491834 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:27:39.498395 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Aug 13 00:27:39.501638 KVP[1700]: KVP starting; pid is:1700 Aug 13 00:27:39.503406 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Aug 13 00:27:39.506030 KVP[1700]: KVP LIC Version: 3.1 Aug 13 00:27:39.506394 kernel: hv_utils: KVP IC version 4.0 Aug 13 00:27:39.507239 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Aug 13 00:27:39.510503 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Aug 13 00:27:39.517452 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Aug 13 00:27:39.522696 systemd[1]: Starting systemd-logind.service - User Login Management... Aug 13 00:27:39.526463 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Aug 13 00:27:39.526847 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Aug 13 00:27:39.528468 systemd[1]: Starting update-engine.service - Update Engine... Aug 13 00:27:39.531850 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Aug 13 00:27:39.539584 google_oslogin_nss_cache[1696]: oslogin_cache_refresh[1696]: Refreshing passwd entry cache Aug 13 00:27:39.539158 oslogin_cache_refresh[1696]: Refreshing passwd entry cache Aug 13 00:27:39.545676 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Aug 13 00:27:39.548801 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Aug 13 00:27:39.548979 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Aug 13 00:27:39.555720 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Aug 13 00:27:39.556337 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Aug 13 00:27:39.564989 jq[1710]: true Aug 13 00:27:39.575182 google_oslogin_nss_cache[1696]: oslogin_cache_refresh[1696]: Failure getting users, quitting Aug 13 00:27:39.575182 google_oslogin_nss_cache[1696]: oslogin_cache_refresh[1696]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Aug 13 00:27:39.575182 google_oslogin_nss_cache[1696]: oslogin_cache_refresh[1696]: Refreshing group entry cache Aug 13 00:27:39.574757 oslogin_cache_refresh[1696]: Failure getting users, quitting Aug 13 00:27:39.574775 oslogin_cache_refresh[1696]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Aug 13 00:27:39.574813 oslogin_cache_refresh[1696]: Refreshing group entry cache Aug 13 00:27:39.580037 extend-filesystems[1695]: Found /dev/nvme0n1p6 Aug 13 00:27:39.597828 google_oslogin_nss_cache[1696]: oslogin_cache_refresh[1696]: Failure getting groups, quitting Aug 13 00:27:39.597828 google_oslogin_nss_cache[1696]: oslogin_cache_refresh[1696]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Aug 13 00:27:39.597179 oslogin_cache_refresh[1696]: Failure getting groups, quitting Aug 13 00:27:39.597188 oslogin_cache_refresh[1696]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Aug 13 00:27:39.599537 (ntainerd)[1720]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Aug 13 00:27:39.599904 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Aug 13 00:27:39.600082 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Aug 13 00:27:39.602667 extend-filesystems[1695]: Found /dev/nvme0n1p9 Aug 13 00:27:39.607053 update_engine[1709]: I20250813 00:27:39.606534 1709 main.cc:92] Flatcar Update Engine starting Aug 13 00:27:39.608892 systemd[1]: motdgen.service: Deactivated successfully. Aug 13 00:27:39.609499 extend-filesystems[1695]: Checking size of /dev/nvme0n1p9 Aug 13 00:27:39.609358 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Aug 13 00:27:39.611260 jq[1723]: true Aug 13 00:27:39.618103 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Aug 13 00:27:39.621958 (chronyd)[1689]: chronyd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Aug 13 00:27:39.630565 chronyd[1751]: chronyd version 4.6.1 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +ASYNCDNS +NTS +SECHASH +IPV6 -DEBUG) Aug 13 00:27:39.637733 extend-filesystems[1695]: Old size kept for /dev/nvme0n1p9 Aug 13 00:27:39.639919 systemd[1]: extend-filesystems.service: Deactivated successfully. Aug 13 00:27:39.640096 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Aug 13 00:27:39.646558 tar[1718]: linux-amd64/LICENSE Aug 13 00:27:39.646558 tar[1718]: linux-amd64/helm Aug 13 00:27:39.653570 chronyd[1751]: Timezone right/UTC failed leap second check, ignoring Aug 13 00:27:39.653744 chronyd[1751]: Loaded seccomp filter (level 2) Aug 13 00:27:39.656320 systemd[1]: Started chronyd.service - NTP client/server. Aug 13 00:27:39.714957 dbus-daemon[1692]: [system] SELinux support is enabled Aug 13 00:27:39.715271 systemd[1]: Started dbus.service - D-Bus System Message Bus. Aug 13 00:27:39.719180 bash[1772]: Updated "/home/core/.ssh/authorized_keys" Aug 13 00:27:39.721593 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Aug 13 00:27:39.724553 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Aug 13 00:27:39.724639 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Aug 13 00:27:39.724671 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Aug 13 00:27:39.725247 update_engine[1709]: I20250813 00:27:39.725116 1709 update_check_scheduler.cc:74] Next update check in 6m54s Aug 13 00:27:39.729405 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Aug 13 00:27:39.729428 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Aug 13 00:27:39.732525 systemd[1]: Started update-engine.service - Update Engine. Aug 13 00:27:39.732655 systemd-logind[1708]: New seat seat0. Aug 13 00:27:39.736711 systemd-logind[1708]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Aug 13 00:27:39.772382 systemd[1]: Started locksmithd.service - Cluster reboot manager. Aug 13 00:27:39.773998 systemd[1]: Started systemd-logind.service - User Login Management. Aug 13 00:27:39.806770 coreos-metadata[1691]: Aug 13 00:27:39.806 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Aug 13 00:27:39.823369 coreos-metadata[1691]: Aug 13 00:27:39.823 INFO Fetch successful Aug 13 00:27:39.823369 coreos-metadata[1691]: Aug 13 00:27:39.823 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Aug 13 00:27:39.830563 coreos-metadata[1691]: Aug 13 00:27:39.830 INFO Fetch successful Aug 13 00:27:39.830563 coreos-metadata[1691]: Aug 13 00:27:39.830 INFO Fetching http://168.63.129.16/machine/e66b5fbf-5a3e-4b2e-9f88-d210b6e2d865/edb96b1d%2D2883%2D4bd6%2Da6d6%2D44b374d4d114.%5Fci%2D4372.1.0%2Da%2D4f0cf7bbd9?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Aug 13 00:27:39.833488 coreos-metadata[1691]: Aug 13 00:27:39.833 INFO Fetch successful Aug 13 00:27:39.833631 coreos-metadata[1691]: Aug 13 00:27:39.833 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Aug 13 00:27:39.851250 coreos-metadata[1691]: Aug 13 00:27:39.850 INFO Fetch successful Aug 13 00:27:39.928582 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Aug 13 00:27:39.942024 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Aug 13 00:27:40.098408 locksmithd[1783]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Aug 13 00:27:40.274564 sshd_keygen[1750]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Aug 13 00:27:40.298881 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Aug 13 00:27:40.303602 systemd[1]: Starting issuegen.service - Generate /run/issue... Aug 13 00:27:40.307502 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Aug 13 00:27:40.329028 systemd[1]: issuegen.service: Deactivated successfully. Aug 13 00:27:40.329189 systemd[1]: Finished issuegen.service - Generate /run/issue. Aug 13 00:27:40.336384 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Aug 13 00:27:40.352997 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Aug 13 00:27:40.363581 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Aug 13 00:27:40.370422 systemd[1]: Started getty@tty1.service - Getty on tty1. Aug 13 00:27:40.373252 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Aug 13 00:27:40.377585 systemd[1]: Reached target getty.target - Login Prompts. Aug 13 00:27:40.471872 tar[1718]: linux-amd64/README.md Aug 13 00:27:40.481808 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Aug 13 00:27:40.496864 containerd[1720]: time="2025-08-13T00:27:40Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Aug 13 00:27:40.498316 containerd[1720]: time="2025-08-13T00:27:40.497634961Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 Aug 13 00:27:40.506654 containerd[1720]: time="2025-08-13T00:27:40.506628381Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="8.167µs" Aug 13 00:27:40.506732 containerd[1720]: time="2025-08-13T00:27:40.506722015Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Aug 13 00:27:40.506772 containerd[1720]: time="2025-08-13T00:27:40.506764631Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Aug 13 00:27:40.506902 containerd[1720]: time="2025-08-13T00:27:40.506894368Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Aug 13 00:27:40.506948 containerd[1720]: time="2025-08-13T00:27:40.506940515Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Aug 13 00:27:40.506991 containerd[1720]: time="2025-08-13T00:27:40.506985391Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507059 containerd[1720]: time="2025-08-13T00:27:40.507050711Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507085 containerd[1720]: time="2025-08-13T00:27:40.507079523Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507320 containerd[1720]: time="2025-08-13T00:27:40.507305791Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507368 containerd[1720]: time="2025-08-13T00:27:40.507360267Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507399 containerd[1720]: time="2025-08-13T00:27:40.507392113Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507429 containerd[1720]: time="2025-08-13T00:27:40.507423282Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507516 containerd[1720]: time="2025-08-13T00:27:40.507509411Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507722 containerd[1720]: time="2025-08-13T00:27:40.507710497Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507774 containerd[1720]: time="2025-08-13T00:27:40.507765195Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Aug 13 00:27:40.507850 containerd[1720]: time="2025-08-13T00:27:40.507796947Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Aug 13 00:27:40.507905 containerd[1720]: time="2025-08-13T00:27:40.507895554Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Aug 13 00:27:40.508206 containerd[1720]: time="2025-08-13T00:27:40.508196962Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Aug 13 00:27:40.508307 containerd[1720]: time="2025-08-13T00:27:40.508298418Z" level=info msg="metadata content store policy set" policy=shared Aug 13 00:27:40.520377 containerd[1720]: time="2025-08-13T00:27:40.520349575Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Aug 13 00:27:40.520454 containerd[1720]: time="2025-08-13T00:27:40.520390241Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Aug 13 00:27:40.520454 containerd[1720]: time="2025-08-13T00:27:40.520402784Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Aug 13 00:27:40.520454 containerd[1720]: time="2025-08-13T00:27:40.520412727Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Aug 13 00:27:40.520454 containerd[1720]: time="2025-08-13T00:27:40.520423236Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Aug 13 00:27:40.520454 containerd[1720]: time="2025-08-13T00:27:40.520432790Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Aug 13 00:27:40.520454 containerd[1720]: time="2025-08-13T00:27:40.520444091Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Aug 13 00:27:40.520454 containerd[1720]: time="2025-08-13T00:27:40.520452979Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Aug 13 00:27:40.520589 containerd[1720]: time="2025-08-13T00:27:40.520462421Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Aug 13 00:27:40.520589 containerd[1720]: time="2025-08-13T00:27:40.520471772Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Aug 13 00:27:40.520589 containerd[1720]: time="2025-08-13T00:27:40.520480926Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Aug 13 00:27:40.520589 containerd[1720]: time="2025-08-13T00:27:40.520498196Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Aug 13 00:27:40.520654 containerd[1720]: time="2025-08-13T00:27:40.520588062Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Aug 13 00:27:40.520654 containerd[1720]: time="2025-08-13T00:27:40.520603042Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Aug 13 00:27:40.520654 containerd[1720]: time="2025-08-13T00:27:40.520616144Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Aug 13 00:27:40.520654 containerd[1720]: time="2025-08-13T00:27:40.520625569Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Aug 13 00:27:40.520654 containerd[1720]: time="2025-08-13T00:27:40.520634709Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Aug 13 00:27:40.520654 containerd[1720]: time="2025-08-13T00:27:40.520644033Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Aug 13 00:27:40.520654 containerd[1720]: time="2025-08-13T00:27:40.520653629Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Aug 13 00:27:40.520773 containerd[1720]: time="2025-08-13T00:27:40.520662601Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Aug 13 00:27:40.520773 containerd[1720]: time="2025-08-13T00:27:40.520671861Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Aug 13 00:27:40.520773 containerd[1720]: time="2025-08-13T00:27:40.520680790Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Aug 13 00:27:40.520773 containerd[1720]: time="2025-08-13T00:27:40.520695434Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Aug 13 00:27:40.520773 containerd[1720]: time="2025-08-13T00:27:40.520752444Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Aug 13 00:27:40.520773 containerd[1720]: time="2025-08-13T00:27:40.520763525Z" level=info msg="Start snapshots syncer" Aug 13 00:27:40.520868 containerd[1720]: time="2025-08-13T00:27:40.520781899Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Aug 13 00:27:40.521848 containerd[1720]: time="2025-08-13T00:27:40.520999913Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Aug 13 00:27:40.521848 containerd[1720]: time="2025-08-13T00:27:40.521042160Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521106001Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521172866Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521188714Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521197113Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521206071Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521216026Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521224921Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521233988Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521257396Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521267335Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521276772Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521312013Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521327648Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Aug 13 00:27:40.522005 containerd[1720]: time="2025-08-13T00:27:40.521335610Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521343557Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521357819Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521367392Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521376877Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521389607Z" level=info msg="runtime interface created" Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521394216Z" level=info msg="created NRI interface" Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521401792Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521410681Z" level=info msg="Connect containerd service" Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521430069Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Aug 13 00:27:40.522252 containerd[1720]: time="2025-08-13T00:27:40.521952200Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 00:27:40.860509 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:27:40.867676 (kubelet)[1857]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 00:27:41.394022 containerd[1720]: time="2025-08-13T00:27:41.393957493Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Aug 13 00:27:41.394022 containerd[1720]: time="2025-08-13T00:27:41.394026695Z" level=info msg=serving... address=/run/containerd/containerd.sock Aug 13 00:27:41.394306 containerd[1720]: time="2025-08-13T00:27:41.394051425Z" level=info msg="Start subscribing containerd event" Aug 13 00:27:41.394306 containerd[1720]: time="2025-08-13T00:27:41.394076332Z" level=info msg="Start recovering state" Aug 13 00:27:41.394306 containerd[1720]: time="2025-08-13T00:27:41.394155867Z" level=info msg="Start event monitor" Aug 13 00:27:41.394306 containerd[1720]: time="2025-08-13T00:27:41.394166249Z" level=info msg="Start cni network conf syncer for default" Aug 13 00:27:41.394306 containerd[1720]: time="2025-08-13T00:27:41.394173205Z" level=info msg="Start streaming server" Aug 13 00:27:41.394306 containerd[1720]: time="2025-08-13T00:27:41.394180701Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Aug 13 00:27:41.394306 containerd[1720]: time="2025-08-13T00:27:41.394187345Z" level=info msg="runtime interface starting up..." Aug 13 00:27:41.394306 containerd[1720]: time="2025-08-13T00:27:41.394193060Z" level=info msg="starting plugins..." Aug 13 00:27:41.394306 containerd[1720]: time="2025-08-13T00:27:41.394203267Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Aug 13 00:27:41.394409 systemd[1]: Started containerd.service - containerd container runtime. Aug 13 00:27:41.397647 containerd[1720]: time="2025-08-13T00:27:41.396482546Z" level=info msg="containerd successfully booted in 0.900012s" Aug 13 00:27:41.397067 systemd[1]: Reached target multi-user.target - Multi-User System. Aug 13 00:27:41.399133 systemd[1]: Startup finished in 3.084s (kernel) + 45.040s (initrd) + 9.861s (userspace) = 57.986s. Aug 13 00:27:41.412020 kubelet[1857]: E0813 00:27:41.411988 1857 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:27:41.418011 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:27:41.418132 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:27:41.418396 systemd[1]: kubelet.service: Consumed 900ms CPU time, 266.5M memory peak. Aug 13 00:27:41.591595 login[1840]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Aug 13 00:27:41.591630 login[1841]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Aug 13 00:27:41.675762 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Aug 13 00:27:41.677381 systemd-logind[1708]: New session 2 of user core. Aug 13 00:27:41.678205 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Aug 13 00:27:41.686350 systemd-logind[1708]: New session 1 of user core. Aug 13 00:27:41.700916 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Aug 13 00:27:41.703366 systemd[1]: Starting user@500.service - User Manager for UID 500... Aug 13 00:27:41.715710 (systemd)[1878]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:27:41.717905 systemd-logind[1708]: New session c1 of user core. Aug 13 00:27:41.763064 waagent[1838]: 2025-08-13T00:27:41.762994Z INFO Daemon Daemon Azure Linux Agent Version: 2.12.0.4 Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.763267Z INFO Daemon Daemon OS: flatcar 4372.1.0 Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.763491Z INFO Daemon Daemon Python: 3.11.12 Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.763748Z INFO Daemon Daemon Run daemon Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.763970Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4372.1.0' Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.764491Z INFO Daemon Daemon Using waagent for provisioning Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.764654Z INFO Daemon Daemon Activate resource disk Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.764867Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.766297Z INFO Daemon Daemon Found device: None Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.766449Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.766660Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.767083Z INFO Daemon Daemon Clean protocol and wireserver endpoint Aug 13 00:27:41.768350 waagent[1838]: 2025-08-13T00:27:41.767224Z INFO Daemon Daemon Running default provisioning handler Aug 13 00:27:41.795556 waagent[1838]: 2025-08-13T00:27:41.795507Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Aug 13 00:27:41.801112 waagent[1838]: 2025-08-13T00:27:41.801073Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Aug 13 00:27:41.806300 waagent[1838]: 2025-08-13T00:27:41.805130Z INFO Daemon Daemon cloud-init is enabled: False Aug 13 00:27:41.807378 waagent[1838]: 2025-08-13T00:27:41.807285Z INFO Daemon Daemon Copying ovf-env.xml Aug 13 00:27:41.867207 waagent[1838]: 2025-08-13T00:27:41.865557Z INFO Daemon Daemon Successfully mounted dvd Aug 13 00:27:41.881510 systemd[1878]: Queued start job for default target default.target. Aug 13 00:27:41.889322 systemd[1878]: Created slice app.slice - User Application Slice. Aug 13 00:27:41.889350 systemd[1878]: Reached target paths.target - Paths. Aug 13 00:27:41.889381 systemd[1878]: Reached target timers.target - Timers. Aug 13 00:27:41.890377 systemd[1878]: Starting dbus.socket - D-Bus User Message Bus Socket... Aug 13 00:27:41.897044 systemd[1878]: Listening on dbus.socket - D-Bus User Message Bus Socket. Aug 13 00:27:41.897894 systemd[1878]: Reached target sockets.target - Sockets. Aug 13 00:27:41.897979 systemd[1878]: Reached target basic.target - Basic System. Aug 13 00:27:41.898006 systemd[1878]: Reached target default.target - Main User Target. Aug 13 00:27:41.898027 systemd[1878]: Startup finished in 174ms. Aug 13 00:27:41.898163 systemd[1]: Started user@500.service - User Manager for UID 500. Aug 13 00:27:41.901465 waagent[1838]: 2025-08-13T00:27:41.901418Z INFO Daemon Daemon Detect protocol endpoint Aug 13 00:27:41.903169 waagent[1838]: 2025-08-13T00:27:41.901587Z INFO Daemon Daemon Clean protocol and wireserver endpoint Aug 13 00:27:41.903169 waagent[1838]: 2025-08-13T00:27:41.901799Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Aug 13 00:27:41.903169 waagent[1838]: 2025-08-13T00:27:41.902434Z INFO Daemon Daemon Test for route to 168.63.129.16 Aug 13 00:27:41.903169 waagent[1838]: 2025-08-13T00:27:41.902579Z INFO Daemon Daemon Route to 168.63.129.16 exists Aug 13 00:27:41.903169 waagent[1838]: 2025-08-13T00:27:41.902824Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Aug 13 00:27:41.907690 systemd[1]: Started session-1.scope - Session 1 of User core. Aug 13 00:27:41.908567 systemd[1]: Started session-2.scope - Session 2 of User core. Aug 13 00:27:41.913041 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Aug 13 00:27:41.920446 waagent[1838]: 2025-08-13T00:27:41.918685Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Aug 13 00:27:41.920446 waagent[1838]: 2025-08-13T00:27:41.918951Z INFO Daemon Daemon Wire protocol version:2012-11-30 Aug 13 00:27:41.920446 waagent[1838]: 2025-08-13T00:27:41.919134Z INFO Daemon Daemon Server preferred version:2015-04-05 Aug 13 00:27:42.040391 waagent[1838]: 2025-08-13T00:27:42.040335Z INFO Daemon Daemon Initializing goal state during protocol detection Aug 13 00:27:42.042646 waagent[1838]: 2025-08-13T00:27:42.040518Z INFO Daemon Daemon Forcing an update of the goal state. Aug 13 00:27:42.045158 waagent[1838]: 2025-08-13T00:27:42.045122Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Aug 13 00:27:42.063569 waagent[1838]: 2025-08-13T00:27:42.063544Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.175 Aug 13 00:27:42.064743 waagent[1838]: 2025-08-13T00:27:42.063997Z INFO Daemon Aug 13 00:27:42.064743 waagent[1838]: 2025-08-13T00:27:42.064527Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: f3a32962-4914-488f-93f7-31c620cc8ec1 eTag: 13320741455721383069 source: Fabric] Aug 13 00:27:42.065038 waagent[1838]: 2025-08-13T00:27:42.064931Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Aug 13 00:27:42.066502 waagent[1838]: 2025-08-13T00:27:42.065212Z INFO Daemon Aug 13 00:27:42.066502 waagent[1838]: 2025-08-13T00:27:42.065385Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Aug 13 00:27:42.077503 waagent[1838]: 2025-08-13T00:27:42.074263Z INFO Daemon Daemon Downloading artifacts profile blob Aug 13 00:27:42.151306 waagent[1838]: 2025-08-13T00:27:42.151243Z INFO Daemon Downloaded certificate {'thumbprint': 'A58FB51B649FE7E2710998E3F10689B9F21DE0A2', 'hasPrivateKey': True} Aug 13 00:27:42.153518 waagent[1838]: 2025-08-13T00:27:42.153487Z INFO Daemon Fetch goal state completed Aug 13 00:27:42.170875 waagent[1838]: 2025-08-13T00:27:42.170846Z INFO Daemon Daemon Starting provisioning Aug 13 00:27:42.171867 waagent[1838]: 2025-08-13T00:27:42.171792Z INFO Daemon Daemon Handle ovf-env.xml. Aug 13 00:27:42.172822 waagent[1838]: 2025-08-13T00:27:42.172758Z INFO Daemon Daemon Set hostname [ci-4372.1.0-a-4f0cf7bbd9] Aug 13 00:27:42.175368 waagent[1838]: 2025-08-13T00:27:42.175332Z INFO Daemon Daemon Publish hostname [ci-4372.1.0-a-4f0cf7bbd9] Aug 13 00:27:42.175933 waagent[1838]: 2025-08-13T00:27:42.175572Z INFO Daemon Daemon Examine /proc/net/route for primary interface Aug 13 00:27:42.175933 waagent[1838]: 2025-08-13T00:27:42.175759Z INFO Daemon Daemon Primary interface is [eth0] Aug 13 00:27:42.181945 systemd-networkd[1364]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 00:27:42.181952 systemd-networkd[1364]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:27:42.181974 systemd-networkd[1364]: eth0: DHCP lease lost Aug 13 00:27:42.182722 waagent[1838]: 2025-08-13T00:27:42.182680Z INFO Daemon Daemon Create user account if not exists Aug 13 00:27:42.183968 waagent[1838]: 2025-08-13T00:27:42.183888Z INFO Daemon Daemon User core already exists, skip useradd Aug 13 00:27:42.186437 waagent[1838]: 2025-08-13T00:27:42.183993Z INFO Daemon Daemon Configure sudoer Aug 13 00:27:42.187675 waagent[1838]: 2025-08-13T00:27:42.187628Z INFO Daemon Daemon Configure sshd Aug 13 00:27:42.192432 waagent[1838]: 2025-08-13T00:27:42.192397Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Aug 13 00:27:42.195670 waagent[1838]: 2025-08-13T00:27:42.192531Z INFO Daemon Daemon Deploy ssh public key. Aug 13 00:27:42.213342 systemd-networkd[1364]: eth0: DHCPv4 address 10.200.8.20/24, gateway 10.200.8.1 acquired from 168.63.129.16 Aug 13 00:27:43.271169 waagent[1838]: 2025-08-13T00:27:43.271121Z INFO Daemon Daemon Provisioning complete Aug 13 00:27:43.282391 waagent[1838]: 2025-08-13T00:27:43.282357Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Aug 13 00:27:43.283522 waagent[1838]: 2025-08-13T00:27:43.283494Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Aug 13 00:27:43.283838 waagent[1838]: 2025-08-13T00:27:43.283814Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.12.0.4 is the most current agent Aug 13 00:27:43.381392 waagent[1927]: 2025-08-13T00:27:43.381327Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.12.0.4) Aug 13 00:27:43.381636 waagent[1927]: 2025-08-13T00:27:43.381415Z INFO ExtHandler ExtHandler OS: flatcar 4372.1.0 Aug 13 00:27:43.381636 waagent[1927]: 2025-08-13T00:27:43.381454Z INFO ExtHandler ExtHandler Python: 3.11.12 Aug 13 00:27:43.381636 waagent[1927]: 2025-08-13T00:27:43.381491Z INFO ExtHandler ExtHandler CPU Arch: x86_64 Aug 13 00:27:43.405025 waagent[1927]: 2025-08-13T00:27:43.404976Z INFO ExtHandler ExtHandler Distro: flatcar-4372.1.0; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.12; Arch: x86_64; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.22.0; Aug 13 00:27:43.405147 waagent[1927]: 2025-08-13T00:27:43.405123Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Aug 13 00:27:43.405200 waagent[1927]: 2025-08-13T00:27:43.405178Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Aug 13 00:27:43.410829 waagent[1927]: 2025-08-13T00:27:43.410781Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Aug 13 00:27:43.422102 waagent[1927]: 2025-08-13T00:27:43.422074Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.175 Aug 13 00:27:43.422450 waagent[1927]: 2025-08-13T00:27:43.422424Z INFO ExtHandler Aug 13 00:27:43.422493 waagent[1927]: 2025-08-13T00:27:43.422479Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 3a45aeef-48ef-4e95-9ba5-13f31bd98223 eTag: 13320741455721383069 source: Fabric] Aug 13 00:27:43.422679 waagent[1927]: 2025-08-13T00:27:43.422658Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Aug 13 00:27:43.422996 waagent[1927]: 2025-08-13T00:27:43.422970Z INFO ExtHandler Aug 13 00:27:43.423032 waagent[1927]: 2025-08-13T00:27:43.423013Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Aug 13 00:27:43.425406 waagent[1927]: 2025-08-13T00:27:43.425379Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Aug 13 00:27:43.485155 waagent[1927]: 2025-08-13T00:27:43.485108Z INFO ExtHandler Downloaded certificate {'thumbprint': 'A58FB51B649FE7E2710998E3F10689B9F21DE0A2', 'hasPrivateKey': True} Aug 13 00:27:43.485503 waagent[1927]: 2025-08-13T00:27:43.485474Z INFO ExtHandler Fetch goal state completed Aug 13 00:27:43.497569 waagent[1927]: 2025-08-13T00:27:43.497524Z INFO ExtHandler ExtHandler OpenSSL version: OpenSSL 3.3.3 11 Feb 2025 (Library: OpenSSL 3.3.3 11 Feb 2025) Aug 13 00:27:43.501278 waagent[1927]: 2025-08-13T00:27:43.501231Z INFO ExtHandler ExtHandler WALinuxAgent-2.12.0.4 running as process 1927 Aug 13 00:27:43.501399 waagent[1927]: 2025-08-13T00:27:43.501359Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Aug 13 00:27:43.501613 waagent[1927]: 2025-08-13T00:27:43.501592Z INFO ExtHandler ExtHandler ******** AutoUpdate.UpdateToLatestVersion is set to False, not processing the operation ******** Aug 13 00:27:43.502554 waagent[1927]: 2025-08-13T00:27:43.502524Z INFO ExtHandler ExtHandler [CGI] Cgroup monitoring is not supported on ['flatcar', '4372.1.0', '', 'Flatcar Container Linux by Kinvolk'] Aug 13 00:27:43.502820 waagent[1927]: 2025-08-13T00:27:43.502795Z INFO ExtHandler ExtHandler [CGI] Agent will reset the quotas in case distro: ['flatcar', '4372.1.0', '', 'Flatcar Container Linux by Kinvolk'] went from supported to unsupported Aug 13 00:27:43.502924 waagent[1927]: 2025-08-13T00:27:43.502905Z INFO ExtHandler ExtHandler [CGI] Agent cgroups enabled: False Aug 13 00:27:43.503286 waagent[1927]: 2025-08-13T00:27:43.503267Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Aug 13 00:27:43.521064 waagent[1927]: 2025-08-13T00:27:43.521044Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Aug 13 00:27:43.521193 waagent[1927]: 2025-08-13T00:27:43.521156Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Aug 13 00:27:43.525885 waagent[1927]: 2025-08-13T00:27:43.525822Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Aug 13 00:27:43.530705 systemd[1]: Reload requested from client PID 1942 ('systemctl') (unit waagent.service)... Aug 13 00:27:43.530716 systemd[1]: Reloading... Aug 13 00:27:43.596309 zram_generator::config[1980]: No configuration found. Aug 13 00:27:43.671727 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:27:43.760363 systemd[1]: Reloading finished in 229 ms. Aug 13 00:27:43.782068 waagent[1927]: 2025-08-13T00:27:43.781971Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Aug 13 00:27:43.782123 waagent[1927]: 2025-08-13T00:27:43.782091Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Aug 13 00:27:43.970318 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#70 cmd 0x4a status: scsi 0x0 srb 0x20 hv 0xc0000001 Aug 13 00:27:44.334917 waagent[1927]: 2025-08-13T00:27:44.334847Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Aug 13 00:27:44.335191 waagent[1927]: 2025-08-13T00:27:44.335162Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: 1. configuration enabled [True], 2. cgroups v1 enabled [False] OR cgroups v2 is in use and v2 resource limiting configuration enabled [False], 3. python supported: [True] Aug 13 00:27:44.335902 waagent[1927]: 2025-08-13T00:27:44.335850Z INFO ExtHandler ExtHandler Starting env monitor service. Aug 13 00:27:44.335961 waagent[1927]: 2025-08-13T00:27:44.335936Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Aug 13 00:27:44.336035 waagent[1927]: 2025-08-13T00:27:44.336003Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Aug 13 00:27:44.336350 waagent[1927]: 2025-08-13T00:27:44.336328Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Aug 13 00:27:44.336584 waagent[1927]: 2025-08-13T00:27:44.336554Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Aug 13 00:27:44.336805 waagent[1927]: 2025-08-13T00:27:44.336785Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Aug 13 00:27:44.336805 waagent[1927]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Aug 13 00:27:44.336805 waagent[1927]: eth0 00000000 0108C80A 0003 0 0 1024 00000000 0 0 0 Aug 13 00:27:44.336805 waagent[1927]: eth0 0008C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Aug 13 00:27:44.336805 waagent[1927]: eth0 0108C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Aug 13 00:27:44.336805 waagent[1927]: eth0 10813FA8 0108C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Aug 13 00:27:44.336805 waagent[1927]: eth0 FEA9FEA9 0108C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Aug 13 00:27:44.336950 waagent[1927]: 2025-08-13T00:27:44.336869Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Aug 13 00:27:44.336950 waagent[1927]: 2025-08-13T00:27:44.336914Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Aug 13 00:27:44.337028 waagent[1927]: 2025-08-13T00:27:44.337008Z INFO EnvHandler ExtHandler Configure routes Aug 13 00:27:44.337073 waagent[1927]: 2025-08-13T00:27:44.337056Z INFO EnvHandler ExtHandler Gateway:None Aug 13 00:27:44.337116 waagent[1927]: 2025-08-13T00:27:44.337094Z INFO EnvHandler ExtHandler Routes:None Aug 13 00:27:44.337515 waagent[1927]: 2025-08-13T00:27:44.337488Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Aug 13 00:27:44.337770 waagent[1927]: 2025-08-13T00:27:44.337735Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Aug 13 00:27:44.338425 waagent[1927]: 2025-08-13T00:27:44.338369Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Aug 13 00:27:44.338466 waagent[1927]: 2025-08-13T00:27:44.338433Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Aug 13 00:27:44.338579 waagent[1927]: 2025-08-13T00:27:44.338562Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Aug 13 00:27:44.352334 waagent[1927]: 2025-08-13T00:27:44.352277Z INFO ExtHandler ExtHandler Aug 13 00:27:44.352396 waagent[1927]: 2025-08-13T00:27:44.352359Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: 8e863b78-6b77-40d1-9dba-881eee5af686 correlation ee7cf0bf-2b90-49fe-ad2e-9bfb3f9f77a2 created: 2025-08-13T00:26:15.345230Z] Aug 13 00:27:44.352634 waagent[1927]: 2025-08-13T00:27:44.352605Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Aug 13 00:27:44.352986 waagent[1927]: 2025-08-13T00:27:44.352960Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 0 ms] Aug 13 00:27:44.377620 waagent[1927]: 2025-08-13T00:27:44.377106Z INFO MonitorHandler ExtHandler Network interfaces: Aug 13 00:27:44.377620 waagent[1927]: Executing ['ip', '-a', '-o', 'link']: Aug 13 00:27:44.377620 waagent[1927]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Aug 13 00:27:44.377620 waagent[1927]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 7c:ed:8d:40:37:32 brd ff:ff:ff:ff:ff:ff\ alias Network Device Aug 13 00:27:44.377620 waagent[1927]: 3: enP30832s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 7c:ed:8d:40:37:32 brd ff:ff:ff:ff:ff:ff\ altname enP30832p0s0 Aug 13 00:27:44.377620 waagent[1927]: Executing ['ip', '-4', '-a', '-o', 'address']: Aug 13 00:27:44.377620 waagent[1927]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Aug 13 00:27:44.377620 waagent[1927]: 2: eth0 inet 10.200.8.20/24 metric 1024 brd 10.200.8.255 scope global eth0\ valid_lft forever preferred_lft forever Aug 13 00:27:44.377620 waagent[1927]: Executing ['ip', '-6', '-a', '-o', 'address']: Aug 13 00:27:44.377620 waagent[1927]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Aug 13 00:27:44.377620 waagent[1927]: 2: eth0 inet6 fe80::7eed:8dff:fe40:3732/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Aug 13 00:27:44.397340 waagent[1927]: 2025-08-13T00:27:44.396992Z WARNING ExtHandler ExtHandler Failed to get firewall packets: 'iptables -w -t security -L OUTPUT --zero OUTPUT -nxv' failed: 2 (iptables v1.8.11 (nf_tables): Illegal option `--numeric' with this command Aug 13 00:27:44.397340 waagent[1927]: Try `iptables -h' or 'iptables --help' for more information.) Aug 13 00:27:44.397683 waagent[1927]: 2025-08-13T00:27:44.397650Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.12.0.4 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: D43FC51B-762A-4831-B5D6-3E3B9316DABF;DroppedPackets: -1;UpdateGSErrors: 0;AutoUpdate: 0;UpdateMode: SelfUpdate;] Aug 13 00:27:44.502905 waagent[1927]: 2025-08-13T00:27:44.502864Z INFO EnvHandler ExtHandler Created firewall rules for the Azure Fabric: Aug 13 00:27:44.502905 waagent[1927]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Aug 13 00:27:44.502905 waagent[1927]: pkts bytes target prot opt in out source destination Aug 13 00:27:44.502905 waagent[1927]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Aug 13 00:27:44.502905 waagent[1927]: pkts bytes target prot opt in out source destination Aug 13 00:27:44.502905 waagent[1927]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Aug 13 00:27:44.502905 waagent[1927]: pkts bytes target prot opt in out source destination Aug 13 00:27:44.502905 waagent[1927]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Aug 13 00:27:44.502905 waagent[1927]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Aug 13 00:27:44.502905 waagent[1927]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Aug 13 00:27:44.505160 waagent[1927]: 2025-08-13T00:27:44.505119Z INFO EnvHandler ExtHandler Current Firewall rules: Aug 13 00:27:44.505160 waagent[1927]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Aug 13 00:27:44.505160 waagent[1927]: pkts bytes target prot opt in out source destination Aug 13 00:27:44.505160 waagent[1927]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Aug 13 00:27:44.505160 waagent[1927]: pkts bytes target prot opt in out source destination Aug 13 00:27:44.505160 waagent[1927]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Aug 13 00:27:44.505160 waagent[1927]: pkts bytes target prot opt in out source destination Aug 13 00:27:44.505160 waagent[1927]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Aug 13 00:27:44.505160 waagent[1927]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Aug 13 00:27:44.505160 waagent[1927]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Aug 13 00:27:51.436113 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Aug 13 00:27:51.437615 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:27:51.940321 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:27:51.947495 (kubelet)[2078]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 00:27:51.983998 kubelet[2078]: E0813 00:27:51.983960 2078 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:27:51.986549 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:27:51.986670 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:27:51.986954 systemd[1]: kubelet.service: Consumed 128ms CPU time, 110.7M memory peak. Aug 13 00:28:02.186048 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Aug 13 00:28:02.187496 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:28:02.343541 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Aug 13 00:28:02.344558 systemd[1]: Started sshd@0-10.200.8.20:22-10.200.16.10:53774.service - OpenSSH per-connection server daemon (10.200.16.10:53774). Aug 13 00:28:02.741218 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:28:02.744035 (kubelet)[2096]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 00:28:02.776534 kubelet[2096]: E0813 00:28:02.776507 2096 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:28:02.778002 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:28:02.778109 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:28:02.778401 systemd[1]: kubelet.service: Consumed 119ms CPU time, 108M memory peak. Aug 13 00:28:03.139953 sshd[2089]: Accepted publickey for core from 10.200.16.10 port 53774 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:28:03.140986 sshd-session[2089]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:28:03.145273 systemd-logind[1708]: New session 3 of user core. Aug 13 00:28:03.147424 systemd[1]: Started session-3.scope - Session 3 of User core. Aug 13 00:28:03.436564 chronyd[1751]: Selected source PHC0 Aug 13 00:28:03.693752 systemd[1]: Started sshd@1-10.200.8.20:22-10.200.16.10:53790.service - OpenSSH per-connection server daemon (10.200.16.10:53790). Aug 13 00:28:04.321156 sshd[2106]: Accepted publickey for core from 10.200.16.10 port 53790 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:28:04.322216 sshd-session[2106]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:28:04.326102 systemd-logind[1708]: New session 4 of user core. Aug 13 00:28:04.334437 systemd[1]: Started session-4.scope - Session 4 of User core. Aug 13 00:28:04.763381 sshd[2108]: Connection closed by 10.200.16.10 port 53790 Aug 13 00:28:04.763874 sshd-session[2106]: pam_unix(sshd:session): session closed for user core Aug 13 00:28:04.766203 systemd[1]: sshd@1-10.200.8.20:22-10.200.16.10:53790.service: Deactivated successfully. Aug 13 00:28:04.767692 systemd[1]: session-4.scope: Deactivated successfully. Aug 13 00:28:04.768793 systemd-logind[1708]: Session 4 logged out. Waiting for processes to exit. Aug 13 00:28:04.769985 systemd-logind[1708]: Removed session 4. Aug 13 00:28:04.888428 systemd[1]: Started sshd@2-10.200.8.20:22-10.200.16.10:53802.service - OpenSSH per-connection server daemon (10.200.16.10:53802). Aug 13 00:28:05.519931 sshd[2114]: Accepted publickey for core from 10.200.16.10 port 53802 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:28:05.520957 sshd-session[2114]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:28:05.524946 systemd-logind[1708]: New session 5 of user core. Aug 13 00:28:05.529448 systemd[1]: Started session-5.scope - Session 5 of User core. Aug 13 00:28:05.959798 sshd[2116]: Connection closed by 10.200.16.10 port 53802 Aug 13 00:28:05.960276 sshd-session[2114]: pam_unix(sshd:session): session closed for user core Aug 13 00:28:05.962806 systemd[1]: sshd@2-10.200.8.20:22-10.200.16.10:53802.service: Deactivated successfully. Aug 13 00:28:05.964266 systemd[1]: session-5.scope: Deactivated successfully. Aug 13 00:28:05.965804 systemd-logind[1708]: Session 5 logged out. Waiting for processes to exit. Aug 13 00:28:05.966625 systemd-logind[1708]: Removed session 5. Aug 13 00:28:06.073947 systemd[1]: Started sshd@3-10.200.8.20:22-10.200.16.10:53814.service - OpenSSH per-connection server daemon (10.200.16.10:53814). Aug 13 00:28:06.710591 sshd[2122]: Accepted publickey for core from 10.200.16.10 port 53814 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:28:06.711636 sshd-session[2122]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:28:06.715856 systemd-logind[1708]: New session 6 of user core. Aug 13 00:28:06.723436 systemd[1]: Started session-6.scope - Session 6 of User core. Aug 13 00:28:07.151930 sshd[2124]: Connection closed by 10.200.16.10 port 53814 Aug 13 00:28:07.152579 sshd-session[2122]: pam_unix(sshd:session): session closed for user core Aug 13 00:28:07.155089 systemd[1]: sshd@3-10.200.8.20:22-10.200.16.10:53814.service: Deactivated successfully. Aug 13 00:28:07.156607 systemd[1]: session-6.scope: Deactivated successfully. Aug 13 00:28:07.158056 systemd-logind[1708]: Session 6 logged out. Waiting for processes to exit. Aug 13 00:28:07.159106 systemd-logind[1708]: Removed session 6. Aug 13 00:28:07.266404 systemd[1]: Started sshd@4-10.200.8.20:22-10.200.16.10:53830.service - OpenSSH per-connection server daemon (10.200.16.10:53830). Aug 13 00:28:07.899720 sshd[2130]: Accepted publickey for core from 10.200.16.10 port 53830 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:28:07.900855 sshd-session[2130]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:28:07.905126 systemd-logind[1708]: New session 7 of user core. Aug 13 00:28:07.911450 systemd[1]: Started session-7.scope - Session 7 of User core. Aug 13 00:28:08.347528 sudo[2133]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Aug 13 00:28:08.347715 sudo[2133]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 00:28:08.377112 sudo[2133]: pam_unix(sudo:session): session closed for user root Aug 13 00:28:08.481284 sshd[2132]: Connection closed by 10.200.16.10 port 53830 Aug 13 00:28:08.481917 sshd-session[2130]: pam_unix(sshd:session): session closed for user core Aug 13 00:28:08.484568 systemd[1]: sshd@4-10.200.8.20:22-10.200.16.10:53830.service: Deactivated successfully. Aug 13 00:28:08.485984 systemd[1]: session-7.scope: Deactivated successfully. Aug 13 00:28:08.487510 systemd-logind[1708]: Session 7 logged out. Waiting for processes to exit. Aug 13 00:28:08.488242 systemd-logind[1708]: Removed session 7. Aug 13 00:28:08.602849 systemd[1]: Started sshd@5-10.200.8.20:22-10.200.16.10:53846.service - OpenSSH per-connection server daemon (10.200.16.10:53846). Aug 13 00:28:09.231342 sshd[2139]: Accepted publickey for core from 10.200.16.10 port 53846 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:28:09.232408 sshd-session[2139]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:28:09.236283 systemd-logind[1708]: New session 8 of user core. Aug 13 00:28:09.242454 systemd[1]: Started session-8.scope - Session 8 of User core. Aug 13 00:28:09.574990 sudo[2143]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Aug 13 00:28:09.575461 sudo[2143]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 00:28:09.581407 sudo[2143]: pam_unix(sudo:session): session closed for user root Aug 13 00:28:09.585004 sudo[2142]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Aug 13 00:28:09.585215 sudo[2142]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 00:28:09.592676 systemd[1]: Starting audit-rules.service - Load Audit Rules... Aug 13 00:28:09.622154 augenrules[2165]: No rules Aug 13 00:28:09.622963 systemd[1]: audit-rules.service: Deactivated successfully. Aug 13 00:28:09.623166 systemd[1]: Finished audit-rules.service - Load Audit Rules. Aug 13 00:28:09.624030 sudo[2142]: pam_unix(sudo:session): session closed for user root Aug 13 00:28:09.730027 sshd[2141]: Connection closed by 10.200.16.10 port 53846 Aug 13 00:28:09.730441 sshd-session[2139]: pam_unix(sshd:session): session closed for user core Aug 13 00:28:09.733073 systemd[1]: sshd@5-10.200.8.20:22-10.200.16.10:53846.service: Deactivated successfully. Aug 13 00:28:09.734526 systemd[1]: session-8.scope: Deactivated successfully. Aug 13 00:28:09.735105 systemd-logind[1708]: Session 8 logged out. Waiting for processes to exit. Aug 13 00:28:09.736143 systemd-logind[1708]: Removed session 8. Aug 13 00:28:09.841517 systemd[1]: Started sshd@6-10.200.8.20:22-10.200.16.10:53862.service - OpenSSH per-connection server daemon (10.200.16.10:53862). Aug 13 00:28:10.475801 sshd[2174]: Accepted publickey for core from 10.200.16.10 port 53862 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:28:10.476860 sshd-session[2174]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:28:10.480848 systemd-logind[1708]: New session 9 of user core. Aug 13 00:28:10.489445 systemd[1]: Started session-9.scope - Session 9 of User core. Aug 13 00:28:10.818227 sudo[2177]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Aug 13 00:28:10.818453 sudo[2177]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 00:28:12.294436 systemd[1]: Starting docker.service - Docker Application Container Engine... Aug 13 00:28:12.307558 (dockerd)[2195]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Aug 13 00:28:12.882281 dockerd[2195]: time="2025-08-13T00:28:12.882223382Z" level=info msg="Starting up" Aug 13 00:28:12.882936 dockerd[2195]: time="2025-08-13T00:28:12.882909125Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Aug 13 00:28:12.883755 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Aug 13 00:28:12.886652 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:28:13.452419 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:28:13.460570 (kubelet)[2223]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 00:28:13.463131 dockerd[2195]: time="2025-08-13T00:28:13.462528973Z" level=info msg="Loading containers: start." Aug 13 00:28:13.492313 kernel: Initializing XFRM netlink socket Aug 13 00:28:13.494605 kubelet[2223]: E0813 00:28:13.494572 2223 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:28:13.496651 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:28:13.496766 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:28:13.497275 systemd[1]: kubelet.service: Consumed 124ms CPU time, 110.4M memory peak. Aug 13 00:28:13.724359 systemd-networkd[1364]: docker0: Link UP Aug 13 00:28:13.746524 dockerd[2195]: time="2025-08-13T00:28:13.746485454Z" level=info msg="Loading containers: done." Aug 13 00:28:13.768856 dockerd[2195]: time="2025-08-13T00:28:13.768826670Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Aug 13 00:28:13.768971 dockerd[2195]: time="2025-08-13T00:28:13.768908204Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 Aug 13 00:28:13.769007 dockerd[2195]: time="2025-08-13T00:28:13.768996584Z" level=info msg="Initializing buildkit" Aug 13 00:28:13.837348 dockerd[2195]: time="2025-08-13T00:28:13.837324113Z" level=info msg="Completed buildkit initialization" Aug 13 00:28:13.843244 dockerd[2195]: time="2025-08-13T00:28:13.843206412Z" level=info msg="Daemon has completed initialization" Aug 13 00:28:13.843347 dockerd[2195]: time="2025-08-13T00:28:13.843274769Z" level=info msg="API listen on /run/docker.sock" Aug 13 00:28:13.843503 systemd[1]: Started docker.service - Docker Application Container Engine. Aug 13 00:28:14.598686 containerd[1720]: time="2025-08-13T00:28:14.598647043Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.3\"" Aug 13 00:28:15.355873 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3556340960.mount: Deactivated successfully. Aug 13 00:28:16.468251 containerd[1720]: time="2025-08-13T00:28:16.468206094Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:16.470431 containerd[1720]: time="2025-08-13T00:28:16.470397456Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.3: active requests=0, bytes read=30078245" Aug 13 00:28:16.473740 containerd[1720]: time="2025-08-13T00:28:16.473711258Z" level=info msg="ImageCreate event name:\"sha256:a92b4b92a991677d355596cc4aa9b0b12cbc38e8cbdc1e476548518ae045bc4a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:16.477301 containerd[1720]: time="2025-08-13T00:28:16.477246559Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:125a8b488def5ea24e2de5682ab1abf063163aae4d89ce21811a45f3ecf23816\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:16.478029 containerd[1720]: time="2025-08-13T00:28:16.477866253Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.3\" with image id \"sha256:a92b4b92a991677d355596cc4aa9b0b12cbc38e8cbdc1e476548518ae045bc4a\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.3\", repo digest \"registry.k8s.io/kube-apiserver@sha256:125a8b488def5ea24e2de5682ab1abf063163aae4d89ce21811a45f3ecf23816\", size \"30075037\" in 1.879186258s" Aug 13 00:28:16.478029 containerd[1720]: time="2025-08-13T00:28:16.477895853Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.3\" returns image reference \"sha256:a92b4b92a991677d355596cc4aa9b0b12cbc38e8cbdc1e476548518ae045bc4a\"" Aug 13 00:28:16.478501 containerd[1720]: time="2025-08-13T00:28:16.478482412Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.3\"" Aug 13 00:28:17.803554 containerd[1720]: time="2025-08-13T00:28:17.803510388Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:17.805641 containerd[1720]: time="2025-08-13T00:28:17.805613210Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.3: active requests=0, bytes read=26019369" Aug 13 00:28:17.807978 containerd[1720]: time="2025-08-13T00:28:17.807953072Z" level=info msg="ImageCreate event name:\"sha256:bf97fadcef43049604abcf0caf4f35229fbee25bd0cdb6fdc1d2bbb4f03d9660\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:17.811437 containerd[1720]: time="2025-08-13T00:28:17.811398190Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:96091626e37c5d5920ee6c3203b783cc01a08f287ec0713aeb7809bb62ccea90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:17.812037 containerd[1720]: time="2025-08-13T00:28:17.811924722Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.3\" with image id \"sha256:bf97fadcef43049604abcf0caf4f35229fbee25bd0cdb6fdc1d2bbb4f03d9660\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.3\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:96091626e37c5d5920ee6c3203b783cc01a08f287ec0713aeb7809bb62ccea90\", size \"27646922\" in 1.333392679s" Aug 13 00:28:17.812037 containerd[1720]: time="2025-08-13T00:28:17.811951325Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.3\" returns image reference \"sha256:bf97fadcef43049604abcf0caf4f35229fbee25bd0cdb6fdc1d2bbb4f03d9660\"" Aug 13 00:28:17.812548 containerd[1720]: time="2025-08-13T00:28:17.812518681Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.3\"" Aug 13 00:28:18.967563 containerd[1720]: time="2025-08-13T00:28:18.967521078Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:18.969744 containerd[1720]: time="2025-08-13T00:28:18.969713575Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.3: active requests=0, bytes read=20155021" Aug 13 00:28:18.972641 containerd[1720]: time="2025-08-13T00:28:18.972603714Z" level=info msg="ImageCreate event name:\"sha256:41376797d5122e388663ab6d0ad583e58cff63e1a0f1eebfb31d615d8f1c1c87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:18.977669 containerd[1720]: time="2025-08-13T00:28:18.977627225Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:f3a2ffdd7483168205236f7762e9a1933f17dd733bc0188b52bddab9c0762868\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:18.978342 containerd[1720]: time="2025-08-13T00:28:18.978210551Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.3\" with image id \"sha256:41376797d5122e388663ab6d0ad583e58cff63e1a0f1eebfb31d615d8f1c1c87\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.3\", repo digest \"registry.k8s.io/kube-scheduler@sha256:f3a2ffdd7483168205236f7762e9a1933f17dd733bc0188b52bddab9c0762868\", size \"21782592\" in 1.165659533s" Aug 13 00:28:18.978342 containerd[1720]: time="2025-08-13T00:28:18.978241757Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.3\" returns image reference \"sha256:41376797d5122e388663ab6d0ad583e58cff63e1a0f1eebfb31d615d8f1c1c87\"" Aug 13 00:28:18.978848 containerd[1720]: time="2025-08-13T00:28:18.978824878Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.3\"" Aug 13 00:28:19.841012 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2074327997.mount: Deactivated successfully. Aug 13 00:28:20.199434 containerd[1720]: time="2025-08-13T00:28:20.199394232Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:20.201618 containerd[1720]: time="2025-08-13T00:28:20.201595439Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.3: active requests=0, bytes read=31892674" Aug 13 00:28:20.204160 containerd[1720]: time="2025-08-13T00:28:20.204139365Z" level=info msg="ImageCreate event name:\"sha256:af855adae796077ff822e22c0102f686b2ca7b7c51948889b1825388eaac9234\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:20.207147 containerd[1720]: time="2025-08-13T00:28:20.207123435Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:c69929cfba9e38305eb1e20ca859aeb90e0d2a7326eab9bb1e8298882fe626cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:20.207421 containerd[1720]: time="2025-08-13T00:28:20.207399114Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.3\" with image id \"sha256:af855adae796077ff822e22c0102f686b2ca7b7c51948889b1825388eaac9234\", repo tag \"registry.k8s.io/kube-proxy:v1.33.3\", repo digest \"registry.k8s.io/kube-proxy@sha256:c69929cfba9e38305eb1e20ca859aeb90e0d2a7326eab9bb1e8298882fe626cd\", size \"31891685\" in 1.228533456s" Aug 13 00:28:20.207624 containerd[1720]: time="2025-08-13T00:28:20.207480118Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.3\" returns image reference \"sha256:af855adae796077ff822e22c0102f686b2ca7b7c51948889b1825388eaac9234\"" Aug 13 00:28:20.207996 containerd[1720]: time="2025-08-13T00:28:20.207975757Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Aug 13 00:28:20.807360 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3691118647.mount: Deactivated successfully. Aug 13 00:28:21.707712 containerd[1720]: time="2025-08-13T00:28:21.707665999Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:21.711501 containerd[1720]: time="2025-08-13T00:28:21.711464104Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942246" Aug 13 00:28:21.714828 containerd[1720]: time="2025-08-13T00:28:21.714787366Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:21.719775 containerd[1720]: time="2025-08-13T00:28:21.719731282Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:21.720511 containerd[1720]: time="2025-08-13T00:28:21.720366758Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.512365652s" Aug 13 00:28:21.720511 containerd[1720]: time="2025-08-13T00:28:21.720396646Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Aug 13 00:28:21.720951 containerd[1720]: time="2025-08-13T00:28:21.720936729Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Aug 13 00:28:22.242220 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1926174748.mount: Deactivated successfully. Aug 13 00:28:22.264973 containerd[1720]: time="2025-08-13T00:28:22.264939396Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 00:28:22.267047 containerd[1720]: time="2025-08-13T00:28:22.267024547Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321146" Aug 13 00:28:22.270664 containerd[1720]: time="2025-08-13T00:28:22.270618170Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 00:28:22.276588 containerd[1720]: time="2025-08-13T00:28:22.276545939Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 00:28:22.277193 containerd[1720]: time="2025-08-13T00:28:22.276906391Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 555.890163ms" Aug 13 00:28:22.277193 containerd[1720]: time="2025-08-13T00:28:22.276935539Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Aug 13 00:28:22.277415 containerd[1720]: time="2025-08-13T00:28:22.277401488Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Aug 13 00:28:22.861854 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3767559626.mount: Deactivated successfully. Aug 13 00:28:23.628225 kernel: hv_balloon: Max. dynamic memory size: 8192 MB Aug 13 00:28:23.686268 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Aug 13 00:28:23.689140 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:28:24.194731 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:28:24.203574 (kubelet)[2597]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 00:28:24.248374 kubelet[2597]: E0813 00:28:24.248340 2597 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:28:24.250960 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:28:24.251194 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:28:24.251754 systemd[1]: kubelet.service: Consumed 138ms CPU time, 108.1M memory peak. Aug 13 00:28:24.530892 containerd[1720]: time="2025-08-13T00:28:24.530815767Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:24.537285 containerd[1720]: time="2025-08-13T00:28:24.537259914Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58247183" Aug 13 00:28:24.540122 containerd[1720]: time="2025-08-13T00:28:24.540083389Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:24.543710 containerd[1720]: time="2025-08-13T00:28:24.543662769Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:24.544622 containerd[1720]: time="2025-08-13T00:28:24.544355555Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 2.266881241s" Aug 13 00:28:24.544622 containerd[1720]: time="2025-08-13T00:28:24.544381424Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Aug 13 00:28:25.424302 update_engine[1709]: I20250813 00:28:25.424239 1709 update_attempter.cc:509] Updating boot flags... Aug 13 00:28:26.800803 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:28:26.801298 systemd[1]: kubelet.service: Consumed 138ms CPU time, 108.1M memory peak. Aug 13 00:28:26.803374 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:28:26.829483 systemd[1]: Reload requested from client PID 2669 ('systemctl') (unit session-9.scope)... Aug 13 00:28:26.829497 systemd[1]: Reloading... Aug 13 00:28:26.915324 zram_generator::config[2718]: No configuration found. Aug 13 00:28:27.030234 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:28:27.118988 systemd[1]: Reloading finished in 289 ms. Aug 13 00:28:27.152633 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Aug 13 00:28:27.152706 systemd[1]: kubelet.service: Failed with result 'signal'. Aug 13 00:28:27.152993 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:28:27.153042 systemd[1]: kubelet.service: Consumed 69ms CPU time, 74.4M memory peak. Aug 13 00:28:27.154262 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:28:27.685275 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:28:27.690641 (kubelet)[2782]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 00:28:27.722827 kubelet[2782]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:28:27.722827 kubelet[2782]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Aug 13 00:28:27.722827 kubelet[2782]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:28:27.723083 kubelet[2782]: I0813 00:28:27.722900 2782 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 00:28:28.198325 kubelet[2782]: I0813 00:28:28.198031 2782 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Aug 13 00:28:28.198325 kubelet[2782]: I0813 00:28:28.198055 2782 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 00:28:28.198526 kubelet[2782]: I0813 00:28:28.198512 2782 server.go:956] "Client rotation is on, will bootstrap in background" Aug 13 00:28:28.223057 kubelet[2782]: E0813 00:28:28.223024 2782 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.200.8.20:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.8.20:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Aug 13 00:28:28.223341 kubelet[2782]: I0813 00:28:28.223325 2782 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 00:28:28.228563 kubelet[2782]: I0813 00:28:28.228550 2782 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Aug 13 00:28:28.232276 kubelet[2782]: I0813 00:28:28.232257 2782 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 00:28:28.232477 kubelet[2782]: I0813 00:28:28.232457 2782 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 00:28:28.232610 kubelet[2782]: I0813 00:28:28.232473 2782 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4372.1.0-a-4f0cf7bbd9","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 00:28:28.232716 kubelet[2782]: I0813 00:28:28.232610 2782 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 00:28:28.232716 kubelet[2782]: I0813 00:28:28.232621 2782 container_manager_linux.go:303] "Creating device plugin manager" Aug 13 00:28:28.233347 kubelet[2782]: I0813 00:28:28.233333 2782 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:28:28.235954 kubelet[2782]: I0813 00:28:28.235782 2782 kubelet.go:480] "Attempting to sync node with API server" Aug 13 00:28:28.235954 kubelet[2782]: I0813 00:28:28.235801 2782 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 00:28:28.235954 kubelet[2782]: I0813 00:28:28.235821 2782 kubelet.go:386] "Adding apiserver pod source" Aug 13 00:28:28.235954 kubelet[2782]: I0813 00:28:28.235835 2782 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 00:28:28.243210 kubelet[2782]: E0813 00:28:28.242859 2782 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.200.8.20:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4372.1.0-a-4f0cf7bbd9&limit=500&resourceVersion=0\": dial tcp 10.200.8.20:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Aug 13 00:28:28.245492 kubelet[2782]: E0813 00:28:28.245465 2782 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.200.8.20:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.8.20:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Aug 13 00:28:28.245565 kubelet[2782]: I0813 00:28:28.245547 2782 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" Aug 13 00:28:28.246345 kubelet[2782]: I0813 00:28:28.246327 2782 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Aug 13 00:28:28.247432 kubelet[2782]: W0813 00:28:28.247420 2782 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Aug 13 00:28:28.249716 kubelet[2782]: I0813 00:28:28.249676 2782 watchdog_linux.go:99] "Systemd watchdog is not enabled" Aug 13 00:28:28.249839 kubelet[2782]: I0813 00:28:28.249833 2782 server.go:1289] "Started kubelet" Aug 13 00:28:28.254468 kubelet[2782]: I0813 00:28:28.254423 2782 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 00:28:28.255263 kubelet[2782]: E0813 00:28:28.253991 2782 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.8.20:6443/api/v1/namespaces/default/events\": dial tcp 10.200.8.20:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4372.1.0-a-4f0cf7bbd9.185b2c0ba1168268 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4372.1.0-a-4f0cf7bbd9,UID:ci-4372.1.0-a-4f0cf7bbd9,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4372.1.0-a-4f0cf7bbd9,},FirstTimestamp:2025-08-13 00:28:28.249809512 +0000 UTC m=+0.556242446,LastTimestamp:2025-08-13 00:28:28.249809512 +0000 UTC m=+0.556242446,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4372.1.0-a-4f0cf7bbd9,}" Aug 13 00:28:28.255414 kubelet[2782]: I0813 00:28:28.255405 2782 server.go:317] "Adding debug handlers to kubelet server" Aug 13 00:28:28.257729 kubelet[2782]: I0813 00:28:28.257397 2782 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 00:28:28.257729 kubelet[2782]: I0813 00:28:28.257689 2782 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 00:28:28.258892 kubelet[2782]: I0813 00:28:28.258870 2782 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 00:28:28.261316 kubelet[2782]: I0813 00:28:28.259765 2782 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 00:28:28.263757 kubelet[2782]: E0813 00:28:28.263739 2782 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" Aug 13 00:28:28.264349 kubelet[2782]: I0813 00:28:28.264338 2782 volume_manager.go:297] "Starting Kubelet Volume Manager" Aug 13 00:28:28.264591 kubelet[2782]: I0813 00:28:28.264581 2782 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Aug 13 00:28:28.264772 kubelet[2782]: I0813 00:28:28.264765 2782 reconciler.go:26] "Reconciler: start to sync state" Aug 13 00:28:28.265366 kubelet[2782]: E0813 00:28:28.265350 2782 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.200.8.20:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.8.20:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Aug 13 00:28:28.266019 kubelet[2782]: I0813 00:28:28.266007 2782 factory.go:223] Registration of the systemd container factory successfully Aug 13 00:28:28.266153 kubelet[2782]: I0813 00:28:28.266143 2782 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 00:28:28.266480 kubelet[2782]: E0813 00:28:28.266468 2782 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 00:28:28.267211 kubelet[2782]: E0813 00:28:28.267188 2782 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.8.20:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4372.1.0-a-4f0cf7bbd9?timeout=10s\": dial tcp 10.200.8.20:6443: connect: connection refused" interval="200ms" Aug 13 00:28:28.267304 kubelet[2782]: I0813 00:28:28.267282 2782 factory.go:223] Registration of the containerd container factory successfully Aug 13 00:28:28.290354 kubelet[2782]: I0813 00:28:28.290343 2782 cpu_manager.go:221] "Starting CPU manager" policy="none" Aug 13 00:28:28.290488 kubelet[2782]: I0813 00:28:28.290480 2782 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Aug 13 00:28:28.290557 kubelet[2782]: I0813 00:28:28.290551 2782 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:28:28.296957 kubelet[2782]: I0813 00:28:28.296945 2782 policy_none.go:49] "None policy: Start" Aug 13 00:28:28.297019 kubelet[2782]: I0813 00:28:28.297015 2782 memory_manager.go:186] "Starting memorymanager" policy="None" Aug 13 00:28:28.297046 kubelet[2782]: I0813 00:28:28.297043 2782 state_mem.go:35] "Initializing new in-memory state store" Aug 13 00:28:28.303575 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Aug 13 00:28:28.308031 kubelet[2782]: I0813 00:28:28.308015 2782 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Aug 13 00:28:28.308981 kubelet[2782]: I0813 00:28:28.308777 2782 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Aug 13 00:28:28.308981 kubelet[2782]: I0813 00:28:28.308798 2782 status_manager.go:230] "Starting to sync pod status with apiserver" Aug 13 00:28:28.308981 kubelet[2782]: I0813 00:28:28.308815 2782 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Aug 13 00:28:28.308981 kubelet[2782]: I0813 00:28:28.308821 2782 kubelet.go:2436] "Starting kubelet main sync loop" Aug 13 00:28:28.308981 kubelet[2782]: E0813 00:28:28.308842 2782 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 00:28:28.313144 kubelet[2782]: E0813 00:28:28.313128 2782 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.200.8.20:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.8.20:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Aug 13 00:28:28.314446 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Aug 13 00:28:28.316735 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Aug 13 00:28:28.320719 kubelet[2782]: E0813 00:28:28.320701 2782 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Aug 13 00:28:28.320831 kubelet[2782]: I0813 00:28:28.320820 2782 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 00:28:28.320856 kubelet[2782]: I0813 00:28:28.320830 2782 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 00:28:28.321347 kubelet[2782]: I0813 00:28:28.321216 2782 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 00:28:28.322475 kubelet[2782]: E0813 00:28:28.322460 2782 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Aug 13 00:28:28.322532 kubelet[2782]: E0813 00:28:28.322494 2782 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" Aug 13 00:28:28.419668 systemd[1]: Created slice kubepods-burstable-podce1c25dd5d25d25ab847f17de2afb9e3.slice - libcontainer container kubepods-burstable-podce1c25dd5d25d25ab847f17de2afb9e3.slice. Aug 13 00:28:28.421916 kubelet[2782]: I0813 00:28:28.421898 2782 kubelet_node_status.go:75] "Attempting to register node" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.422139 kubelet[2782]: E0813 00:28:28.422123 2782 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.8.20:6443/api/v1/nodes\": dial tcp 10.200.8.20:6443: connect: connection refused" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.427891 kubelet[2782]: E0813 00:28:28.427835 2782 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.434063 systemd[1]: Created slice kubepods-burstable-poda5a9c80705d9a3c6281288b129fe0cfd.slice - libcontainer container kubepods-burstable-poda5a9c80705d9a3c6281288b129fe0cfd.slice. Aug 13 00:28:28.435875 kubelet[2782]: E0813 00:28:28.435756 2782 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.437631 systemd[1]: Created slice kubepods-burstable-pod6cdcd462b85f754bcc3195ea6126eab4.slice - libcontainer container kubepods-burstable-pod6cdcd462b85f754bcc3195ea6126eab4.slice. Aug 13 00:28:28.438926 kubelet[2782]: E0813 00:28:28.438909 2782 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.466309 kubelet[2782]: I0813 00:28:28.466116 2782 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-ca-certs\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.466309 kubelet[2782]: I0813 00:28:28.466147 2782 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-flexvolume-dir\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.466309 kubelet[2782]: I0813 00:28:28.466163 2782 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.466309 kubelet[2782]: I0813 00:28:28.466179 2782 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6cdcd462b85f754bcc3195ea6126eab4-kubeconfig\") pod \"kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"6cdcd462b85f754bcc3195ea6126eab4\") " pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.466309 kubelet[2782]: I0813 00:28:28.466200 2782 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ce1c25dd5d25d25ab847f17de2afb9e3-k8s-certs\") pod \"kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"ce1c25dd5d25d25ab847f17de2afb9e3\") " pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.466455 kubelet[2782]: I0813 00:28:28.466223 2782 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ce1c25dd5d25d25ab847f17de2afb9e3-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"ce1c25dd5d25d25ab847f17de2afb9e3\") " pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.466455 kubelet[2782]: I0813 00:28:28.466237 2782 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-k8s-certs\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.466455 kubelet[2782]: I0813 00:28:28.466250 2782 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-kubeconfig\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.466455 kubelet[2782]: I0813 00:28:28.466263 2782 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ce1c25dd5d25d25ab847f17de2afb9e3-ca-certs\") pod \"kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"ce1c25dd5d25d25ab847f17de2afb9e3\") " pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.468475 kubelet[2782]: E0813 00:28:28.468454 2782 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.8.20:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4372.1.0-a-4f0cf7bbd9?timeout=10s\": dial tcp 10.200.8.20:6443: connect: connection refused" interval="400ms" Aug 13 00:28:28.623918 kubelet[2782]: I0813 00:28:28.623880 2782 kubelet_node_status.go:75] "Attempting to register node" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.624197 kubelet[2782]: E0813 00:28:28.624175 2782 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.8.20:6443/api/v1/nodes\": dial tcp 10.200.8.20:6443: connect: connection refused" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:28.728760 containerd[1720]: time="2025-08-13T00:28:28.728681820Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9,Uid:ce1c25dd5d25d25ab847f17de2afb9e3,Namespace:kube-system,Attempt:0,}" Aug 13 00:28:28.737097 containerd[1720]: time="2025-08-13T00:28:28.737072546Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9,Uid:a5a9c80705d9a3c6281288b129fe0cfd,Namespace:kube-system,Attempt:0,}" Aug 13 00:28:28.739733 containerd[1720]: time="2025-08-13T00:28:28.739709535Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9,Uid:6cdcd462b85f754bcc3195ea6126eab4,Namespace:kube-system,Attempt:0,}" Aug 13 00:28:28.793967 containerd[1720]: time="2025-08-13T00:28:28.793933289Z" level=info msg="connecting to shim 4cef4e2a1aeea6d678af33dc7ef02fa1a312b0833fc9c5047801dce9b8e544d8" address="unix:///run/containerd/s/1842a5d4abd944150ecf09df166accd13e5e5b98d0c11732d33ca737c8b64090" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:28:28.813224 containerd[1720]: time="2025-08-13T00:28:28.813173343Z" level=info msg="connecting to shim a86f3e33b4efe390509c6cb7f5ed38ac79cb2a87a5771f1e48a345cb95ec0f07" address="unix:///run/containerd/s/c05a0b3594ac62596664bda8c4b41cebb5ccd59f1aaafa4367cdcf1cc38e166b" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:28:28.820462 systemd[1]: Started cri-containerd-4cef4e2a1aeea6d678af33dc7ef02fa1a312b0833fc9c5047801dce9b8e544d8.scope - libcontainer container 4cef4e2a1aeea6d678af33dc7ef02fa1a312b0833fc9c5047801dce9b8e544d8. Aug 13 00:28:28.839872 containerd[1720]: time="2025-08-13T00:28:28.839837969Z" level=info msg="connecting to shim e1d6007a1dbeeecf04195cfdf4cca750dd025444bbd1c29abf34c6bbb9d23e7c" address="unix:///run/containerd/s/3c0d51aebc2df4cb8f656dfd984c8dd581ae40d58a3b087bd2fb6e6f1725ef22" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:28:28.842445 systemd[1]: Started cri-containerd-a86f3e33b4efe390509c6cb7f5ed38ac79cb2a87a5771f1e48a345cb95ec0f07.scope - libcontainer container a86f3e33b4efe390509c6cb7f5ed38ac79cb2a87a5771f1e48a345cb95ec0f07. Aug 13 00:28:28.870120 kubelet[2782]: E0813 00:28:28.869091 2782 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.8.20:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4372.1.0-a-4f0cf7bbd9?timeout=10s\": dial tcp 10.200.8.20:6443: connect: connection refused" interval="800ms" Aug 13 00:28:28.872495 systemd[1]: Started cri-containerd-e1d6007a1dbeeecf04195cfdf4cca750dd025444bbd1c29abf34c6bbb9d23e7c.scope - libcontainer container e1d6007a1dbeeecf04195cfdf4cca750dd025444bbd1c29abf34c6bbb9d23e7c. Aug 13 00:28:28.905970 containerd[1720]: time="2025-08-13T00:28:28.905949477Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9,Uid:6cdcd462b85f754bcc3195ea6126eab4,Namespace:kube-system,Attempt:0,} returns sandbox id \"a86f3e33b4efe390509c6cb7f5ed38ac79cb2a87a5771f1e48a345cb95ec0f07\"" Aug 13 00:28:28.909675 containerd[1720]: time="2025-08-13T00:28:28.909647223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9,Uid:ce1c25dd5d25d25ab847f17de2afb9e3,Namespace:kube-system,Attempt:0,} returns sandbox id \"4cef4e2a1aeea6d678af33dc7ef02fa1a312b0833fc9c5047801dce9b8e544d8\"" Aug 13 00:28:28.913333 containerd[1720]: time="2025-08-13T00:28:28.913111406Z" level=info msg="CreateContainer within sandbox \"a86f3e33b4efe390509c6cb7f5ed38ac79cb2a87a5771f1e48a345cb95ec0f07\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Aug 13 00:28:28.918322 containerd[1720]: time="2025-08-13T00:28:28.918281352Z" level=info msg="CreateContainer within sandbox \"4cef4e2a1aeea6d678af33dc7ef02fa1a312b0833fc9c5047801dce9b8e544d8\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Aug 13 00:28:28.936649 containerd[1720]: time="2025-08-13T00:28:28.936625799Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9,Uid:a5a9c80705d9a3c6281288b129fe0cfd,Namespace:kube-system,Attempt:0,} returns sandbox id \"e1d6007a1dbeeecf04195cfdf4cca750dd025444bbd1c29abf34c6bbb9d23e7c\"" Aug 13 00:28:28.945599 containerd[1720]: time="2025-08-13T00:28:28.945572979Z" level=info msg="CreateContainer within sandbox \"e1d6007a1dbeeecf04195cfdf4cca750dd025444bbd1c29abf34c6bbb9d23e7c\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Aug 13 00:28:28.947215 containerd[1720]: time="2025-08-13T00:28:28.947194196Z" level=info msg="Container d8247f196649287b9e92e847422f2d5d49a3af1cf1b7cc46090e02e4f2497c82: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:28:28.951635 containerd[1720]: time="2025-08-13T00:28:28.951612610Z" level=info msg="Container 0ba922613a9d094e2dfc69e537690fe503d127b27ff68135d0582cc71a0c07b7: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:28:28.976055 containerd[1720]: time="2025-08-13T00:28:28.976032720Z" level=info msg="Container d99a76c1cce1efd113e8a0cdab262c410e57de3ab028da98ed0cb3746b6a11d1: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:28:28.987598 containerd[1720]: time="2025-08-13T00:28:28.987534877Z" level=info msg="CreateContainer within sandbox \"a86f3e33b4efe390509c6cb7f5ed38ac79cb2a87a5771f1e48a345cb95ec0f07\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"d8247f196649287b9e92e847422f2d5d49a3af1cf1b7cc46090e02e4f2497c82\"" Aug 13 00:28:28.988307 containerd[1720]: time="2025-08-13T00:28:28.988245737Z" level=info msg="CreateContainer within sandbox \"4cef4e2a1aeea6d678af33dc7ef02fa1a312b0833fc9c5047801dce9b8e544d8\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"0ba922613a9d094e2dfc69e537690fe503d127b27ff68135d0582cc71a0c07b7\"" Aug 13 00:28:28.988475 containerd[1720]: time="2025-08-13T00:28:28.988459173Z" level=info msg="StartContainer for \"d8247f196649287b9e92e847422f2d5d49a3af1cf1b7cc46090e02e4f2497c82\"" Aug 13 00:28:28.989123 containerd[1720]: time="2025-08-13T00:28:28.989093164Z" level=info msg="connecting to shim d8247f196649287b9e92e847422f2d5d49a3af1cf1b7cc46090e02e4f2497c82" address="unix:///run/containerd/s/c05a0b3594ac62596664bda8c4b41cebb5ccd59f1aaafa4367cdcf1cc38e166b" protocol=ttrpc version=3 Aug 13 00:28:28.989590 containerd[1720]: time="2025-08-13T00:28:28.989565278Z" level=info msg="StartContainer for \"0ba922613a9d094e2dfc69e537690fe503d127b27ff68135d0582cc71a0c07b7\"" Aug 13 00:28:28.990409 containerd[1720]: time="2025-08-13T00:28:28.990383069Z" level=info msg="connecting to shim 0ba922613a9d094e2dfc69e537690fe503d127b27ff68135d0582cc71a0c07b7" address="unix:///run/containerd/s/1842a5d4abd944150ecf09df166accd13e5e5b98d0c11732d33ca737c8b64090" protocol=ttrpc version=3 Aug 13 00:28:29.001794 containerd[1720]: time="2025-08-13T00:28:29.001716502Z" level=info msg="CreateContainer within sandbox \"e1d6007a1dbeeecf04195cfdf4cca750dd025444bbd1c29abf34c6bbb9d23e7c\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"d99a76c1cce1efd113e8a0cdab262c410e57de3ab028da98ed0cb3746b6a11d1\"" Aug 13 00:28:29.002711 containerd[1720]: time="2025-08-13T00:28:29.002671608Z" level=info msg="StartContainer for \"d99a76c1cce1efd113e8a0cdab262c410e57de3ab028da98ed0cb3746b6a11d1\"" Aug 13 00:28:29.003614 containerd[1720]: time="2025-08-13T00:28:29.003589257Z" level=info msg="connecting to shim d99a76c1cce1efd113e8a0cdab262c410e57de3ab028da98ed0cb3746b6a11d1" address="unix:///run/containerd/s/3c0d51aebc2df4cb8f656dfd984c8dd581ae40d58a3b087bd2fb6e6f1725ef22" protocol=ttrpc version=3 Aug 13 00:28:29.004458 systemd[1]: Started cri-containerd-d8247f196649287b9e92e847422f2d5d49a3af1cf1b7cc46090e02e4f2497c82.scope - libcontainer container d8247f196649287b9e92e847422f2d5d49a3af1cf1b7cc46090e02e4f2497c82. Aug 13 00:28:29.007534 systemd[1]: Started cri-containerd-0ba922613a9d094e2dfc69e537690fe503d127b27ff68135d0582cc71a0c07b7.scope - libcontainer container 0ba922613a9d094e2dfc69e537690fe503d127b27ff68135d0582cc71a0c07b7. Aug 13 00:28:29.025935 kubelet[2782]: I0813 00:28:29.025918 2782 kubelet_node_status.go:75] "Attempting to register node" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:29.026242 kubelet[2782]: E0813 00:28:29.026220 2782 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.8.20:6443/api/v1/nodes\": dial tcp 10.200.8.20:6443: connect: connection refused" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:29.028431 systemd[1]: Started cri-containerd-d99a76c1cce1efd113e8a0cdab262c410e57de3ab028da98ed0cb3746b6a11d1.scope - libcontainer container d99a76c1cce1efd113e8a0cdab262c410e57de3ab028da98ed0cb3746b6a11d1. Aug 13 00:28:29.092009 containerd[1720]: time="2025-08-13T00:28:29.091924622Z" level=info msg="StartContainer for \"0ba922613a9d094e2dfc69e537690fe503d127b27ff68135d0582cc71a0c07b7\" returns successfully" Aug 13 00:28:29.095309 containerd[1720]: time="2025-08-13T00:28:29.095268874Z" level=info msg="StartContainer for \"d8247f196649287b9e92e847422f2d5d49a3af1cf1b7cc46090e02e4f2497c82\" returns successfully" Aug 13 00:28:29.114916 containerd[1720]: time="2025-08-13T00:28:29.114785569Z" level=info msg="StartContainer for \"d99a76c1cce1efd113e8a0cdab262c410e57de3ab028da98ed0cb3746b6a11d1\" returns successfully" Aug 13 00:28:29.317863 kubelet[2782]: E0813 00:28:29.317784 2782 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:29.320432 kubelet[2782]: E0813 00:28:29.320414 2782 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:29.321924 kubelet[2782]: E0813 00:28:29.321900 2782 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:29.828264 kubelet[2782]: I0813 00:28:29.828238 2782 kubelet_node_status.go:75] "Attempting to register node" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:30.325499 kubelet[2782]: E0813 00:28:30.325403 2782 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:30.325841 kubelet[2782]: E0813 00:28:30.325704 2782 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4372.1.0-a-4f0cf7bbd9\" not found" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.197900 kubelet[2782]: E0813 00:28:31.197864 2782 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4372.1.0-a-4f0cf7bbd9\" not found" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.244608 kubelet[2782]: I0813 00:28:31.244572 2782 apiserver.go:52] "Watching apiserver" Aug 13 00:28:31.264934 kubelet[2782]: I0813 00:28:31.264908 2782 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Aug 13 00:28:31.281165 kubelet[2782]: I0813 00:28:31.280980 2782 kubelet_node_status.go:78] "Successfully registered node" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.367621 kubelet[2782]: I0813 00:28:31.367599 2782 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.371154 kubelet[2782]: E0813 00:28:31.371126 2782 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.371154 kubelet[2782]: I0813 00:28:31.371144 2782 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.372309 kubelet[2782]: E0813 00:28:31.372260 2782 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.372309 kubelet[2782]: I0813 00:28:31.372278 2782 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.373267 kubelet[2782]: E0813 00:28:31.373244 2782 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.621147 kubelet[2782]: I0813 00:28:31.621074 2782 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:31.622701 kubelet[2782]: E0813 00:28:31.622679 2782 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:32.313258 kubelet[2782]: I0813 00:28:32.313229 2782 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:32.320127 kubelet[2782]: I0813 00:28:32.320089 2782 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:28:32.772252 kubelet[2782]: I0813 00:28:32.772195 2782 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:32.776481 kubelet[2782]: I0813 00:28:32.776359 2782 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:28:33.132452 systemd[1]: Reload requested from client PID 3056 ('systemctl') (unit session-9.scope)... Aug 13 00:28:33.132466 systemd[1]: Reloading... Aug 13 00:28:33.204313 zram_generator::config[3102]: No configuration found. Aug 13 00:28:33.283093 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:28:33.380547 systemd[1]: Reloading finished in 247 ms. Aug 13 00:28:33.411104 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:28:33.427463 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 00:28:33.427758 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:28:33.427836 systemd[1]: kubelet.service: Consumed 836ms CPU time, 129.3M memory peak. Aug 13 00:28:33.430116 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:28:33.931269 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:28:33.937669 (kubelet)[3169]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 00:28:33.972977 kubelet[3169]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:28:33.972977 kubelet[3169]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Aug 13 00:28:33.972977 kubelet[3169]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:28:33.972977 kubelet[3169]: I0813 00:28:33.972835 3169 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 00:28:33.979927 kubelet[3169]: I0813 00:28:33.979903 3169 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Aug 13 00:28:33.979927 kubelet[3169]: I0813 00:28:33.979922 3169 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 00:28:33.980195 kubelet[3169]: I0813 00:28:33.980081 3169 server.go:956] "Client rotation is on, will bootstrap in background" Aug 13 00:28:33.981060 kubelet[3169]: I0813 00:28:33.981043 3169 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Aug 13 00:28:33.982904 kubelet[3169]: I0813 00:28:33.982474 3169 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 00:28:33.986326 kubelet[3169]: I0813 00:28:33.986278 3169 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Aug 13 00:28:33.988513 kubelet[3169]: I0813 00:28:33.988454 3169 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 00:28:33.988667 kubelet[3169]: I0813 00:28:33.988623 3169 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 00:28:33.990666 kubelet[3169]: I0813 00:28:33.988647 3169 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4372.1.0-a-4f0cf7bbd9","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 00:28:33.990666 kubelet[3169]: I0813 00:28:33.989655 3169 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 00:28:33.990666 kubelet[3169]: I0813 00:28:33.989670 3169 container_manager_linux.go:303] "Creating device plugin manager" Aug 13 00:28:33.990666 kubelet[3169]: I0813 00:28:33.989716 3169 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:28:33.990666 kubelet[3169]: I0813 00:28:33.989853 3169 kubelet.go:480] "Attempting to sync node with API server" Aug 13 00:28:33.990891 kubelet[3169]: I0813 00:28:33.989864 3169 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 00:28:33.990891 kubelet[3169]: I0813 00:28:33.989883 3169 kubelet.go:386] "Adding apiserver pod source" Aug 13 00:28:33.990891 kubelet[3169]: I0813 00:28:33.989897 3169 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 00:28:34.001231 kubelet[3169]: I0813 00:28:34.000655 3169 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" Aug 13 00:28:34.001231 kubelet[3169]: I0813 00:28:34.001081 3169 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Aug 13 00:28:34.003112 kubelet[3169]: I0813 00:28:34.003016 3169 watchdog_linux.go:99] "Systemd watchdog is not enabled" Aug 13 00:28:34.003112 kubelet[3169]: I0813 00:28:34.003056 3169 server.go:1289] "Started kubelet" Aug 13 00:28:34.003682 kubelet[3169]: I0813 00:28:34.003438 3169 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 00:28:34.003768 kubelet[3169]: I0813 00:28:34.003355 3169 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 00:28:34.004745 kubelet[3169]: I0813 00:28:34.004731 3169 server.go:317] "Adding debug handlers to kubelet server" Aug 13 00:28:34.005332 kubelet[3169]: I0813 00:28:34.004770 3169 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 00:28:34.006849 kubelet[3169]: I0813 00:28:34.006806 3169 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 00:28:34.008435 kubelet[3169]: I0813 00:28:34.008219 3169 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 00:28:34.010756 kubelet[3169]: I0813 00:28:34.010709 3169 volume_manager.go:297] "Starting Kubelet Volume Manager" Aug 13 00:28:34.010817 kubelet[3169]: I0813 00:28:34.010774 3169 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Aug 13 00:28:34.010970 kubelet[3169]: I0813 00:28:34.010864 3169 reconciler.go:26] "Reconciler: start to sync state" Aug 13 00:28:34.012587 kubelet[3169]: E0813 00:28:34.012570 3169 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 00:28:34.014429 kubelet[3169]: I0813 00:28:34.014411 3169 factory.go:223] Registration of the containerd container factory successfully Aug 13 00:28:34.014429 kubelet[3169]: I0813 00:28:34.014428 3169 factory.go:223] Registration of the systemd container factory successfully Aug 13 00:28:34.014528 kubelet[3169]: I0813 00:28:34.014496 3169 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 00:28:34.020186 kubelet[3169]: I0813 00:28:34.020154 3169 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Aug 13 00:28:34.020992 kubelet[3169]: I0813 00:28:34.020968 3169 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Aug 13 00:28:34.020992 kubelet[3169]: I0813 00:28:34.020986 3169 status_manager.go:230] "Starting to sync pod status with apiserver" Aug 13 00:28:34.021083 kubelet[3169]: I0813 00:28:34.021001 3169 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Aug 13 00:28:34.021083 kubelet[3169]: I0813 00:28:34.021008 3169 kubelet.go:2436] "Starting kubelet main sync loop" Aug 13 00:28:34.021083 kubelet[3169]: E0813 00:28:34.021036 3169 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 00:28:34.060756 kubelet[3169]: I0813 00:28:34.060735 3169 cpu_manager.go:221] "Starting CPU manager" policy="none" Aug 13 00:28:34.060756 kubelet[3169]: I0813 00:28:34.060746 3169 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Aug 13 00:28:34.060871 kubelet[3169]: I0813 00:28:34.060774 3169 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:28:34.060871 kubelet[3169]: I0813 00:28:34.060867 3169 state_mem.go:88] "Updated default CPUSet" cpuSet="" Aug 13 00:28:34.060916 kubelet[3169]: I0813 00:28:34.060875 3169 state_mem.go:96] "Updated CPUSet assignments" assignments={} Aug 13 00:28:34.060916 kubelet[3169]: I0813 00:28:34.060891 3169 policy_none.go:49] "None policy: Start" Aug 13 00:28:34.060916 kubelet[3169]: I0813 00:28:34.060899 3169 memory_manager.go:186] "Starting memorymanager" policy="None" Aug 13 00:28:34.060916 kubelet[3169]: I0813 00:28:34.060908 3169 state_mem.go:35] "Initializing new in-memory state store" Aug 13 00:28:34.061002 kubelet[3169]: I0813 00:28:34.060980 3169 state_mem.go:75] "Updated machine memory state" Aug 13 00:28:34.063737 kubelet[3169]: E0813 00:28:34.063723 3169 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Aug 13 00:28:34.063918 kubelet[3169]: I0813 00:28:34.063910 3169 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 00:28:34.064113 kubelet[3169]: I0813 00:28:34.063963 3169 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 00:28:34.066649 kubelet[3169]: I0813 00:28:34.064556 3169 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 00:28:34.068431 kubelet[3169]: E0813 00:28:34.068223 3169 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Aug 13 00:28:34.121891 kubelet[3169]: I0813 00:28:34.121876 3169 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.122534 kubelet[3169]: I0813 00:28:34.121876 3169 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.122534 kubelet[3169]: I0813 00:28:34.121952 3169 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.135373 kubelet[3169]: I0813 00:28:34.135360 3169 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:28:34.135640 kubelet[3169]: I0813 00:28:34.135480 3169 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:28:34.135735 kubelet[3169]: E0813 00:28:34.135724 3169 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" already exists" pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.135831 kubelet[3169]: I0813 00:28:34.135498 3169 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:28:34.135831 kubelet[3169]: E0813 00:28:34.135809 3169 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9\" already exists" pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.169697 kubelet[3169]: I0813 00:28:34.169674 3169 kubelet_node_status.go:75] "Attempting to register node" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.180516 kubelet[3169]: I0813 00:28:34.180501 3169 kubelet_node_status.go:124] "Node was previously registered" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.180749 kubelet[3169]: I0813 00:28:34.180561 3169 kubelet_node_status.go:78] "Successfully registered node" node="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.312438 kubelet[3169]: I0813 00:28:34.312412 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ce1c25dd5d25d25ab847f17de2afb9e3-ca-certs\") pod \"kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"ce1c25dd5d25d25ab847f17de2afb9e3\") " pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.312601 kubelet[3169]: I0813 00:28:34.312442 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ce1c25dd5d25d25ab847f17de2afb9e3-k8s-certs\") pod \"kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"ce1c25dd5d25d25ab847f17de2afb9e3\") " pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.312601 kubelet[3169]: I0813 00:28:34.312461 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ce1c25dd5d25d25ab847f17de2afb9e3-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"ce1c25dd5d25d25ab847f17de2afb9e3\") " pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.312601 kubelet[3169]: I0813 00:28:34.312499 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-flexvolume-dir\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.312601 kubelet[3169]: I0813 00:28:34.312521 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6cdcd462b85f754bcc3195ea6126eab4-kubeconfig\") pod \"kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"6cdcd462b85f754bcc3195ea6126eab4\") " pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.312601 kubelet[3169]: I0813 00:28:34.312540 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-ca-certs\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.312758 kubelet[3169]: I0813 00:28:34.312562 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-k8s-certs\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.312758 kubelet[3169]: I0813 00:28:34.312586 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-kubeconfig\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.312758 kubelet[3169]: I0813 00:28:34.312602 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a5a9c80705d9a3c6281288b129fe0cfd-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9\" (UID: \"a5a9c80705d9a3c6281288b129fe0cfd\") " pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:34.993334 kubelet[3169]: I0813 00:28:34.993303 3169 apiserver.go:52] "Watching apiserver" Aug 13 00:28:35.011760 kubelet[3169]: I0813 00:28:35.011739 3169 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Aug 13 00:28:35.052151 kubelet[3169]: I0813 00:28:35.052062 3169 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:35.052536 kubelet[3169]: I0813 00:28:35.052524 3169 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:35.061270 kubelet[3169]: I0813 00:28:35.061020 3169 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:28:35.061270 kubelet[3169]: E0813 00:28:35.061068 3169 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9\" already exists" pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:35.061883 kubelet[3169]: I0813 00:28:35.061868 3169 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:28:35.062009 kubelet[3169]: E0813 00:28:35.062000 3169 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9\" already exists" pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:28:35.074748 kubelet[3169]: I0813 00:28:35.074510 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4372.1.0-a-4f0cf7bbd9" podStartSLOduration=1.07449781 podStartE2EDuration="1.07449781s" podCreationTimestamp="2025-08-13 00:28:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:28:35.067255923 +0000 UTC m=+1.126287033" watchObservedRunningTime="2025-08-13 00:28:35.07449781 +0000 UTC m=+1.133528892" Aug 13 00:28:35.074748 kubelet[3169]: I0813 00:28:35.074639 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4372.1.0-a-4f0cf7bbd9" podStartSLOduration=3.074633156 podStartE2EDuration="3.074633156s" podCreationTimestamp="2025-08-13 00:28:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:28:35.07437912 +0000 UTC m=+1.133410207" watchObservedRunningTime="2025-08-13 00:28:35.074633156 +0000 UTC m=+1.133664242" Aug 13 00:28:35.081706 kubelet[3169]: I0813 00:28:35.081650 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4372.1.0-a-4f0cf7bbd9" podStartSLOduration=3.081642152 podStartE2EDuration="3.081642152s" podCreationTimestamp="2025-08-13 00:28:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:28:35.081520503 +0000 UTC m=+1.140551590" watchObservedRunningTime="2025-08-13 00:28:35.081642152 +0000 UTC m=+1.140673236" Aug 13 00:28:39.795239 kubelet[3169]: I0813 00:28:39.795212 3169 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Aug 13 00:28:39.795620 containerd[1720]: time="2025-08-13T00:28:39.795572807Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Aug 13 00:28:39.795789 kubelet[3169]: I0813 00:28:39.795725 3169 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Aug 13 00:28:40.668423 systemd[1]: Created slice kubepods-besteffort-pod1a466bbf_15fb_4f41_83c8_026cbaac72c5.slice - libcontainer container kubepods-besteffort-pod1a466bbf_15fb_4f41_83c8_026cbaac72c5.slice. Aug 13 00:28:40.755779 kubelet[3169]: I0813 00:28:40.755736 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/1a466bbf-15fb-4f41-83c8-026cbaac72c5-kube-proxy\") pod \"kube-proxy-cbdhq\" (UID: \"1a466bbf-15fb-4f41-83c8-026cbaac72c5\") " pod="kube-system/kube-proxy-cbdhq" Aug 13 00:28:40.755779 kubelet[3169]: I0813 00:28:40.755780 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gchvk\" (UniqueName: \"kubernetes.io/projected/1a466bbf-15fb-4f41-83c8-026cbaac72c5-kube-api-access-gchvk\") pod \"kube-proxy-cbdhq\" (UID: \"1a466bbf-15fb-4f41-83c8-026cbaac72c5\") " pod="kube-system/kube-proxy-cbdhq" Aug 13 00:28:40.755779 kubelet[3169]: I0813 00:28:40.755797 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1a466bbf-15fb-4f41-83c8-026cbaac72c5-xtables-lock\") pod \"kube-proxy-cbdhq\" (UID: \"1a466bbf-15fb-4f41-83c8-026cbaac72c5\") " pod="kube-system/kube-proxy-cbdhq" Aug 13 00:28:40.755954 kubelet[3169]: I0813 00:28:40.755811 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1a466bbf-15fb-4f41-83c8-026cbaac72c5-lib-modules\") pod \"kube-proxy-cbdhq\" (UID: \"1a466bbf-15fb-4f41-83c8-026cbaac72c5\") " pod="kube-system/kube-proxy-cbdhq" Aug 13 00:28:40.933322 systemd[1]: Created slice kubepods-besteffort-podec08783b_03e9_4b73_a49f_1e7542a36236.slice - libcontainer container kubepods-besteffort-podec08783b_03e9_4b73_a49f_1e7542a36236.slice. Aug 13 00:28:40.957397 kubelet[3169]: I0813 00:28:40.957368 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcbqv\" (UniqueName: \"kubernetes.io/projected/ec08783b-03e9-4b73-a49f-1e7542a36236-kube-api-access-zcbqv\") pod \"tigera-operator-747864d56d-8vqqz\" (UID: \"ec08783b-03e9-4b73-a49f-1e7542a36236\") " pod="tigera-operator/tigera-operator-747864d56d-8vqqz" Aug 13 00:28:40.957656 kubelet[3169]: I0813 00:28:40.957404 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/ec08783b-03e9-4b73-a49f-1e7542a36236-var-lib-calico\") pod \"tigera-operator-747864d56d-8vqqz\" (UID: \"ec08783b-03e9-4b73-a49f-1e7542a36236\") " pod="tigera-operator/tigera-operator-747864d56d-8vqqz" Aug 13 00:28:40.974845 containerd[1720]: time="2025-08-13T00:28:40.974808764Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-cbdhq,Uid:1a466bbf-15fb-4f41-83c8-026cbaac72c5,Namespace:kube-system,Attempt:0,}" Aug 13 00:28:41.018927 containerd[1720]: time="2025-08-13T00:28:41.018899819Z" level=info msg="connecting to shim 130823005660e0cadda1f209bb06f0e47f7751e979e2dae44302785c4f89cc66" address="unix:///run/containerd/s/30a0e402f7eeba160f89ca55ef14cbf39e1d16404cc6e5f6c6b085eb704fa03b" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:28:41.038445 systemd[1]: Started cri-containerd-130823005660e0cadda1f209bb06f0e47f7751e979e2dae44302785c4f89cc66.scope - libcontainer container 130823005660e0cadda1f209bb06f0e47f7751e979e2dae44302785c4f89cc66. Aug 13 00:28:41.062558 containerd[1720]: time="2025-08-13T00:28:41.062493345Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-cbdhq,Uid:1a466bbf-15fb-4f41-83c8-026cbaac72c5,Namespace:kube-system,Attempt:0,} returns sandbox id \"130823005660e0cadda1f209bb06f0e47f7751e979e2dae44302785c4f89cc66\"" Aug 13 00:28:41.073020 containerd[1720]: time="2025-08-13T00:28:41.073002576Z" level=info msg="CreateContainer within sandbox \"130823005660e0cadda1f209bb06f0e47f7751e979e2dae44302785c4f89cc66\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Aug 13 00:28:41.102865 containerd[1720]: time="2025-08-13T00:28:41.102821354Z" level=info msg="Container 8aecbc3562608935ea8e48acafdf01b876bfc0cb9dfe179cc980b9a760fd1eba: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:28:41.121238 containerd[1720]: time="2025-08-13T00:28:41.121203040Z" level=info msg="CreateContainer within sandbox \"130823005660e0cadda1f209bb06f0e47f7751e979e2dae44302785c4f89cc66\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"8aecbc3562608935ea8e48acafdf01b876bfc0cb9dfe179cc980b9a760fd1eba\"" Aug 13 00:28:41.122320 containerd[1720]: time="2025-08-13T00:28:41.121649108Z" level=info msg="StartContainer for \"8aecbc3562608935ea8e48acafdf01b876bfc0cb9dfe179cc980b9a760fd1eba\"" Aug 13 00:28:41.123379 containerd[1720]: time="2025-08-13T00:28:41.123353424Z" level=info msg="connecting to shim 8aecbc3562608935ea8e48acafdf01b876bfc0cb9dfe179cc980b9a760fd1eba" address="unix:///run/containerd/s/30a0e402f7eeba160f89ca55ef14cbf39e1d16404cc6e5f6c6b085eb704fa03b" protocol=ttrpc version=3 Aug 13 00:28:41.140397 systemd[1]: Started cri-containerd-8aecbc3562608935ea8e48acafdf01b876bfc0cb9dfe179cc980b9a760fd1eba.scope - libcontainer container 8aecbc3562608935ea8e48acafdf01b876bfc0cb9dfe179cc980b9a760fd1eba. Aug 13 00:28:41.167168 containerd[1720]: time="2025-08-13T00:28:41.167144255Z" level=info msg="StartContainer for \"8aecbc3562608935ea8e48acafdf01b876bfc0cb9dfe179cc980b9a760fd1eba\" returns successfully" Aug 13 00:28:41.236571 containerd[1720]: time="2025-08-13T00:28:41.236546635Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-747864d56d-8vqqz,Uid:ec08783b-03e9-4b73-a49f-1e7542a36236,Namespace:tigera-operator,Attempt:0,}" Aug 13 00:28:41.272761 containerd[1720]: time="2025-08-13T00:28:41.272725958Z" level=info msg="connecting to shim c6116e221cedab3bb90a4f806ca254cac420652ced65544bf76e51eaba13be4a" address="unix:///run/containerd/s/4bcade42057b8f1346e0d5aeb6b13a8819865f554da6437164c9deb5d184e5a3" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:28:41.292416 systemd[1]: Started cri-containerd-c6116e221cedab3bb90a4f806ca254cac420652ced65544bf76e51eaba13be4a.scope - libcontainer container c6116e221cedab3bb90a4f806ca254cac420652ced65544bf76e51eaba13be4a. Aug 13 00:28:41.326470 containerd[1720]: time="2025-08-13T00:28:41.326447972Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-747864d56d-8vqqz,Uid:ec08783b-03e9-4b73-a49f-1e7542a36236,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"c6116e221cedab3bb90a4f806ca254cac420652ced65544bf76e51eaba13be4a\"" Aug 13 00:28:41.327730 containerd[1720]: time="2025-08-13T00:28:41.327670387Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\"" Aug 13 00:28:42.076317 kubelet[3169]: I0813 00:28:42.075935 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-cbdhq" podStartSLOduration=2.075920023 podStartE2EDuration="2.075920023s" podCreationTimestamp="2025-08-13 00:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:28:42.075774707 +0000 UTC m=+8.134805792" watchObservedRunningTime="2025-08-13 00:28:42.075920023 +0000 UTC m=+8.134951110" Aug 13 00:28:42.751128 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3078464760.mount: Deactivated successfully. Aug 13 00:28:43.146482 containerd[1720]: time="2025-08-13T00:28:43.146394945Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:43.148565 containerd[1720]: time="2025-08-13T00:28:43.148533474Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.3: active requests=0, bytes read=25056543" Aug 13 00:28:43.151420 containerd[1720]: time="2025-08-13T00:28:43.151392178Z" level=info msg="ImageCreate event name:\"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:43.156165 containerd[1720]: time="2025-08-13T00:28:43.156116917Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:43.156459 containerd[1720]: time="2025-08-13T00:28:43.156438402Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.3\" with image id \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\", repo tag \"quay.io/tigera/operator:v1.38.3\", repo digest \"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\", size \"25052538\" in 1.828739541s" Aug 13 00:28:43.156498 containerd[1720]: time="2025-08-13T00:28:43.156466802Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\" returns image reference \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\"" Aug 13 00:28:43.162951 containerd[1720]: time="2025-08-13T00:28:43.162923435Z" level=info msg="CreateContainer within sandbox \"c6116e221cedab3bb90a4f806ca254cac420652ced65544bf76e51eaba13be4a\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Aug 13 00:28:43.185728 containerd[1720]: time="2025-08-13T00:28:43.184390822Z" level=info msg="Container 5bb42b0d6d7d64518ba598d73f91c1277c2b46b43fa2415cdf830e66c8b78ba2: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:28:43.198091 containerd[1720]: time="2025-08-13T00:28:43.198067525Z" level=info msg="CreateContainer within sandbox \"c6116e221cedab3bb90a4f806ca254cac420652ced65544bf76e51eaba13be4a\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"5bb42b0d6d7d64518ba598d73f91c1277c2b46b43fa2415cdf830e66c8b78ba2\"" Aug 13 00:28:43.198432 containerd[1720]: time="2025-08-13T00:28:43.198411381Z" level=info msg="StartContainer for \"5bb42b0d6d7d64518ba598d73f91c1277c2b46b43fa2415cdf830e66c8b78ba2\"" Aug 13 00:28:43.199329 containerd[1720]: time="2025-08-13T00:28:43.199308306Z" level=info msg="connecting to shim 5bb42b0d6d7d64518ba598d73f91c1277c2b46b43fa2415cdf830e66c8b78ba2" address="unix:///run/containerd/s/4bcade42057b8f1346e0d5aeb6b13a8819865f554da6437164c9deb5d184e5a3" protocol=ttrpc version=3 Aug 13 00:28:43.219450 systemd[1]: Started cri-containerd-5bb42b0d6d7d64518ba598d73f91c1277c2b46b43fa2415cdf830e66c8b78ba2.scope - libcontainer container 5bb42b0d6d7d64518ba598d73f91c1277c2b46b43fa2415cdf830e66c8b78ba2. Aug 13 00:28:43.244953 containerd[1720]: time="2025-08-13T00:28:43.244796830Z" level=info msg="StartContainer for \"5bb42b0d6d7d64518ba598d73f91c1277c2b46b43fa2415cdf830e66c8b78ba2\" returns successfully" Aug 13 00:28:45.409715 kubelet[3169]: I0813 00:28:45.409660 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-747864d56d-8vqqz" podStartSLOduration=3.5798325970000002 podStartE2EDuration="5.4096416s" podCreationTimestamp="2025-08-13 00:28:40 +0000 UTC" firstStartedPulling="2025-08-13 00:28:41.327283556 +0000 UTC m=+7.386314643" lastFinishedPulling="2025-08-13 00:28:43.157092564 +0000 UTC m=+9.216123646" observedRunningTime="2025-08-13 00:28:44.079228652 +0000 UTC m=+10.138259737" watchObservedRunningTime="2025-08-13 00:28:45.4096416 +0000 UTC m=+11.468672692" Aug 13 00:28:48.624463 sudo[2177]: pam_unix(sudo:session): session closed for user root Aug 13 00:28:48.728017 sshd[2176]: Connection closed by 10.200.16.10 port 53862 Aug 13 00:28:48.729267 sshd-session[2174]: pam_unix(sshd:session): session closed for user core Aug 13 00:28:48.733117 systemd-logind[1708]: Session 9 logged out. Waiting for processes to exit. Aug 13 00:28:48.734837 systemd[1]: sshd@6-10.200.8.20:22-10.200.16.10:53862.service: Deactivated successfully. Aug 13 00:28:48.739552 systemd[1]: session-9.scope: Deactivated successfully. Aug 13 00:28:48.739739 systemd[1]: session-9.scope: Consumed 3.396s CPU time, 231.7M memory peak. Aug 13 00:28:48.743170 systemd-logind[1708]: Removed session 9. Aug 13 00:28:53.178468 systemd[1]: Created slice kubepods-besteffort-pode80f19e4_0aea_4613_ac00_65120748d8e4.slice - libcontainer container kubepods-besteffort-pode80f19e4_0aea_4613_ac00_65120748d8e4.slice. Aug 13 00:28:53.235992 kubelet[3169]: I0813 00:28:53.235941 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e80f19e4-0aea-4613-ac00-65120748d8e4-tigera-ca-bundle\") pod \"calico-typha-658b76bf5b-9wx5r\" (UID: \"e80f19e4-0aea-4613-ac00-65120748d8e4\") " pod="calico-system/calico-typha-658b76bf5b-9wx5r" Aug 13 00:28:53.235992 kubelet[3169]: I0813 00:28:53.235974 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/e80f19e4-0aea-4613-ac00-65120748d8e4-typha-certs\") pod \"calico-typha-658b76bf5b-9wx5r\" (UID: \"e80f19e4-0aea-4613-ac00-65120748d8e4\") " pod="calico-system/calico-typha-658b76bf5b-9wx5r" Aug 13 00:28:53.235992 kubelet[3169]: I0813 00:28:53.235995 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgm79\" (UniqueName: \"kubernetes.io/projected/e80f19e4-0aea-4613-ac00-65120748d8e4-kube-api-access-hgm79\") pod \"calico-typha-658b76bf5b-9wx5r\" (UID: \"e80f19e4-0aea-4613-ac00-65120748d8e4\") " pod="calico-system/calico-typha-658b76bf5b-9wx5r" Aug 13 00:28:53.487134 containerd[1720]: time="2025-08-13T00:28:53.487107448Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-658b76bf5b-9wx5r,Uid:e80f19e4-0aea-4613-ac00-65120748d8e4,Namespace:calico-system,Attempt:0,}" Aug 13 00:28:53.556079 containerd[1720]: time="2025-08-13T00:28:53.555968383Z" level=info msg="connecting to shim 5762273d447720091720b906213ea2a5cc0ea18deeda7bf3430f6e59991700dc" address="unix:///run/containerd/s/b274f104e21fb218cc08ec6ed14d28470ae715e2f5a792443394c4976b919c5d" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:28:53.580626 systemd[1]: Created slice kubepods-besteffort-pod96326c11_f4ca_491f_97e3_ab82bfe78410.slice - libcontainer container kubepods-besteffort-pod96326c11_f4ca_491f_97e3_ab82bfe78410.slice. Aug 13 00:28:53.594581 systemd[1]: Started cri-containerd-5762273d447720091720b906213ea2a5cc0ea18deeda7bf3430f6e59991700dc.scope - libcontainer container 5762273d447720091720b906213ea2a5cc0ea18deeda7bf3430f6e59991700dc. Aug 13 00:28:53.636119 containerd[1720]: time="2025-08-13T00:28:53.636020302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-658b76bf5b-9wx5r,Uid:e80f19e4-0aea-4613-ac00-65120748d8e4,Namespace:calico-system,Attempt:0,} returns sandbox id \"5762273d447720091720b906213ea2a5cc0ea18deeda7bf3430f6e59991700dc\"" Aug 13 00:28:53.638930 kubelet[3169]: I0813 00:28:53.638619 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96326c11-f4ca-491f-97e3-ab82bfe78410-tigera-ca-bundle\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.638930 kubelet[3169]: I0813 00:28:53.638645 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/96326c11-f4ca-491f-97e3-ab82bfe78410-var-run-calico\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.638930 kubelet[3169]: I0813 00:28:53.638662 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvmzr\" (UniqueName: \"kubernetes.io/projected/96326c11-f4ca-491f-97e3-ab82bfe78410-kube-api-access-fvmzr\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.638930 kubelet[3169]: I0813 00:28:53.638707 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/96326c11-f4ca-491f-97e3-ab82bfe78410-cni-bin-dir\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.638930 kubelet[3169]: I0813 00:28:53.638722 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/96326c11-f4ca-491f-97e3-ab82bfe78410-policysync\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.639095 kubelet[3169]: I0813 00:28:53.638737 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/96326c11-f4ca-491f-97e3-ab82bfe78410-xtables-lock\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.639095 kubelet[3169]: I0813 00:28:53.638784 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/96326c11-f4ca-491f-97e3-ab82bfe78410-flexvol-driver-host\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.639095 kubelet[3169]: I0813 00:28:53.638799 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/96326c11-f4ca-491f-97e3-ab82bfe78410-lib-modules\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.639095 kubelet[3169]: I0813 00:28:53.638815 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/96326c11-f4ca-491f-97e3-ab82bfe78410-cni-net-dir\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.639095 kubelet[3169]: I0813 00:28:53.638849 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/96326c11-f4ca-491f-97e3-ab82bfe78410-cni-log-dir\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.639194 kubelet[3169]: I0813 00:28:53.638863 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/96326c11-f4ca-491f-97e3-ab82bfe78410-var-lib-calico\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.639194 kubelet[3169]: I0813 00:28:53.638880 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/96326c11-f4ca-491f-97e3-ab82bfe78410-node-certs\") pod \"calico-node-6mkd5\" (UID: \"96326c11-f4ca-491f-97e3-ab82bfe78410\") " pod="calico-system/calico-node-6mkd5" Aug 13 00:28:53.641250 containerd[1720]: time="2025-08-13T00:28:53.640425576Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\"" Aug 13 00:28:53.740613 kubelet[3169]: E0813 00:28:53.740558 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.741026 kubelet[3169]: W0813 00:28:53.741013 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.741087 kubelet[3169]: E0813 00:28:53.741076 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.741399 kubelet[3169]: E0813 00:28:53.741389 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.741469 kubelet[3169]: W0813 00:28:53.741461 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.741521 kubelet[3169]: E0813 00:28:53.741514 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.741682 kubelet[3169]: E0813 00:28:53.741675 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.741725 kubelet[3169]: W0813 00:28:53.741719 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.741768 kubelet[3169]: E0813 00:28:53.741752 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.741953 kubelet[3169]: E0813 00:28:53.741947 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.742040 kubelet[3169]: W0813 00:28:53.742032 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.742078 kubelet[3169]: E0813 00:28:53.742072 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.744371 kubelet[3169]: E0813 00:28:53.744356 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.744449 kubelet[3169]: W0813 00:28:53.744441 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.745314 kubelet[3169]: E0813 00:28:53.744486 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.745498 kubelet[3169]: E0813 00:28:53.745490 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.745546 kubelet[3169]: W0813 00:28:53.745539 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.745583 kubelet[3169]: E0813 00:28:53.745577 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.757320 kubelet[3169]: E0813 00:28:53.757284 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.757320 kubelet[3169]: W0813 00:28:53.757317 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.757401 kubelet[3169]: E0813 00:28:53.757329 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.856664 kubelet[3169]: E0813 00:28:53.856623 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-ptpbv" podUID="a62490c7-241c-47fd-bc6c-9b6d06fb91b7" Aug 13 00:28:53.883344 containerd[1720]: time="2025-08-13T00:28:53.883319821Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6mkd5,Uid:96326c11-f4ca-491f-97e3-ab82bfe78410,Namespace:calico-system,Attempt:0,}" Aug 13 00:28:53.936032 kubelet[3169]: E0813 00:28:53.936015 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.936032 kubelet[3169]: W0813 00:28:53.936031 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.936265 kubelet[3169]: E0813 00:28:53.936045 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.936395 kubelet[3169]: E0813 00:28:53.936326 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.936395 kubelet[3169]: W0813 00:28:53.936334 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.936395 kubelet[3169]: E0813 00:28:53.936345 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.936574 kubelet[3169]: E0813 00:28:53.936463 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.936574 kubelet[3169]: W0813 00:28:53.936470 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.936574 kubelet[3169]: E0813 00:28:53.936477 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.936732 kubelet[3169]: E0813 00:28:53.936624 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.936732 kubelet[3169]: W0813 00:28:53.936630 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.936732 kubelet[3169]: E0813 00:28:53.936637 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.937441 kubelet[3169]: E0813 00:28:53.937423 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.937441 kubelet[3169]: W0813 00:28:53.937441 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.937534 kubelet[3169]: E0813 00:28:53.937454 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.937559 kubelet[3169]: E0813 00:28:53.937556 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.937584 kubelet[3169]: W0813 00:28:53.937561 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.937584 kubelet[3169]: E0813 00:28:53.937567 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.938064 kubelet[3169]: E0813 00:28:53.937649 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.938064 kubelet[3169]: W0813 00:28:53.937655 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.938064 kubelet[3169]: E0813 00:28:53.937660 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.938064 kubelet[3169]: E0813 00:28:53.937741 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.938064 kubelet[3169]: W0813 00:28:53.937746 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.938064 kubelet[3169]: E0813 00:28:53.937751 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.938064 kubelet[3169]: E0813 00:28:53.937839 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.938064 kubelet[3169]: W0813 00:28:53.937844 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.938064 kubelet[3169]: E0813 00:28:53.937850 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.938064 kubelet[3169]: E0813 00:28:53.937930 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.938833 kubelet[3169]: W0813 00:28:53.937937 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.938833 kubelet[3169]: E0813 00:28:53.937943 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.938833 kubelet[3169]: E0813 00:28:53.938034 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.938833 kubelet[3169]: W0813 00:28:53.938040 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.938833 kubelet[3169]: E0813 00:28:53.938046 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.938833 kubelet[3169]: E0813 00:28:53.938122 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.938833 kubelet[3169]: W0813 00:28:53.938126 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.938833 kubelet[3169]: E0813 00:28:53.938131 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.938833 kubelet[3169]: E0813 00:28:53.938214 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.938833 kubelet[3169]: W0813 00:28:53.938218 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.939024 kubelet[3169]: E0813 00:28:53.938224 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.939024 kubelet[3169]: E0813 00:28:53.938338 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.939024 kubelet[3169]: W0813 00:28:53.938344 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.939024 kubelet[3169]: E0813 00:28:53.938350 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.939024 kubelet[3169]: E0813 00:28:53.938429 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.939024 kubelet[3169]: W0813 00:28:53.938434 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.939024 kubelet[3169]: E0813 00:28:53.938440 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.939024 kubelet[3169]: E0813 00:28:53.938523 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.939024 kubelet[3169]: W0813 00:28:53.938528 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.939024 kubelet[3169]: E0813 00:28:53.938533 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.939208 kubelet[3169]: E0813 00:28:53.938630 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.939208 kubelet[3169]: W0813 00:28:53.938635 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.939208 kubelet[3169]: E0813 00:28:53.938641 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.939208 kubelet[3169]: E0813 00:28:53.938712 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.939208 kubelet[3169]: W0813 00:28:53.938718 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.939208 kubelet[3169]: E0813 00:28:53.938723 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.939208 kubelet[3169]: E0813 00:28:53.938798 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.939208 kubelet[3169]: W0813 00:28:53.938802 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.939208 kubelet[3169]: E0813 00:28:53.938807 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.939208 kubelet[3169]: E0813 00:28:53.938882 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.940583 kubelet[3169]: W0813 00:28:53.938886 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.940583 kubelet[3169]: E0813 00:28:53.938892 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.941224 kubelet[3169]: E0813 00:28:53.941178 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.941224 kubelet[3169]: W0813 00:28:53.941192 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.941224 kubelet[3169]: E0813 00:28:53.941203 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.941476 kubelet[3169]: I0813 00:28:53.941422 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a62490c7-241c-47fd-bc6c-9b6d06fb91b7-kubelet-dir\") pod \"csi-node-driver-ptpbv\" (UID: \"a62490c7-241c-47fd-bc6c-9b6d06fb91b7\") " pod="calico-system/csi-node-driver-ptpbv" Aug 13 00:28:53.942155 kubelet[3169]: E0813 00:28:53.941989 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.942155 kubelet[3169]: W0813 00:28:53.942003 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.942155 kubelet[3169]: E0813 00:28:53.942016 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.942338 kubelet[3169]: E0813 00:28:53.942277 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.942338 kubelet[3169]: W0813 00:28:53.942299 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.942338 kubelet[3169]: I0813 00:28:53.942042 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a62490c7-241c-47fd-bc6c-9b6d06fb91b7-registration-dir\") pod \"csi-node-driver-ptpbv\" (UID: \"a62490c7-241c-47fd-bc6c-9b6d06fb91b7\") " pod="calico-system/csi-node-driver-ptpbv" Aug 13 00:28:53.943080 kubelet[3169]: E0813 00:28:53.942421 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.943080 kubelet[3169]: E0813 00:28:53.943047 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.943080 kubelet[3169]: W0813 00:28:53.943057 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.943080 kubelet[3169]: E0813 00:28:53.943068 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.944808 kubelet[3169]: E0813 00:28:53.944436 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.944808 kubelet[3169]: W0813 00:28:53.944456 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.944808 kubelet[3169]: E0813 00:28:53.944469 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.944808 kubelet[3169]: I0813 00:28:53.944626 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/a62490c7-241c-47fd-bc6c-9b6d06fb91b7-varrun\") pod \"csi-node-driver-ptpbv\" (UID: \"a62490c7-241c-47fd-bc6c-9b6d06fb91b7\") " pod="calico-system/csi-node-driver-ptpbv" Aug 13 00:28:53.944808 kubelet[3169]: E0813 00:28:53.944700 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.944808 kubelet[3169]: W0813 00:28:53.944708 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.944808 kubelet[3169]: E0813 00:28:53.944718 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.945515 kubelet[3169]: E0813 00:28:53.945095 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.945515 kubelet[3169]: W0813 00:28:53.945105 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.945515 kubelet[3169]: E0813 00:28:53.945116 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.945515 kubelet[3169]: E0813 00:28:53.945253 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.945515 kubelet[3169]: W0813 00:28:53.945258 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.945515 kubelet[3169]: E0813 00:28:53.945265 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.945515 kubelet[3169]: I0813 00:28:53.945282 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bgl9\" (UniqueName: \"kubernetes.io/projected/a62490c7-241c-47fd-bc6c-9b6d06fb91b7-kube-api-access-5bgl9\") pod \"csi-node-driver-ptpbv\" (UID: \"a62490c7-241c-47fd-bc6c-9b6d06fb91b7\") " pod="calico-system/csi-node-driver-ptpbv" Aug 13 00:28:53.945515 kubelet[3169]: E0813 00:28:53.945441 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.945515 kubelet[3169]: W0813 00:28:53.945448 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.945756 kubelet[3169]: E0813 00:28:53.945456 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.945971 kubelet[3169]: E0813 00:28:53.945868 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.945971 kubelet[3169]: W0813 00:28:53.945881 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.945971 kubelet[3169]: E0813 00:28:53.945892 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.946463 kubelet[3169]: I0813 00:28:53.946313 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a62490c7-241c-47fd-bc6c-9b6d06fb91b7-socket-dir\") pod \"csi-node-driver-ptpbv\" (UID: \"a62490c7-241c-47fd-bc6c-9b6d06fb91b7\") " pod="calico-system/csi-node-driver-ptpbv" Aug 13 00:28:53.946463 kubelet[3169]: E0813 00:28:53.946370 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.946463 kubelet[3169]: W0813 00:28:53.946378 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.946463 kubelet[3169]: E0813 00:28:53.946388 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.946824 kubelet[3169]: E0813 00:28:53.946781 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.946824 kubelet[3169]: W0813 00:28:53.946792 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.946824 kubelet[3169]: E0813 00:28:53.946802 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.947660 kubelet[3169]: E0813 00:28:53.947495 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.947660 kubelet[3169]: W0813 00:28:53.947509 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.947660 kubelet[3169]: E0813 00:28:53.947521 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.947938 kubelet[3169]: E0813 00:28:53.947784 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.947938 kubelet[3169]: W0813 00:28:53.947795 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.947938 kubelet[3169]: E0813 00:28:53.947806 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.948221 kubelet[3169]: E0813 00:28:53.948179 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:53.948221 kubelet[3169]: W0813 00:28:53.948190 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:53.948221 kubelet[3169]: E0813 00:28:53.948202 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:53.955316 containerd[1720]: time="2025-08-13T00:28:53.955087046Z" level=info msg="connecting to shim c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4" address="unix:///run/containerd/s/5c173d3927e4e19eb263c99381ddee7ad8a6c5ea45d5176f2a82cc7983b8a868" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:28:53.977612 systemd[1]: Started cri-containerd-c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4.scope - libcontainer container c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4. Aug 13 00:28:54.001894 containerd[1720]: time="2025-08-13T00:28:54.001840486Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6mkd5,Uid:96326c11-f4ca-491f-97e3-ab82bfe78410,Namespace:calico-system,Attempt:0,} returns sandbox id \"c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4\"" Aug 13 00:28:54.047405 kubelet[3169]: E0813 00:28:54.047388 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.047405 kubelet[3169]: W0813 00:28:54.047401 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.047540 kubelet[3169]: E0813 00:28:54.047414 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.047540 kubelet[3169]: E0813 00:28:54.047526 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.047540 kubelet[3169]: W0813 00:28:54.047531 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.047540 kubelet[3169]: E0813 00:28:54.047538 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.047655 kubelet[3169]: E0813 00:28:54.047643 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.047655 kubelet[3169]: W0813 00:28:54.047649 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.047701 kubelet[3169]: E0813 00:28:54.047655 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.047760 kubelet[3169]: E0813 00:28:54.047749 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.047760 kubelet[3169]: W0813 00:28:54.047756 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.047825 kubelet[3169]: E0813 00:28:54.047763 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.047850 kubelet[3169]: E0813 00:28:54.047846 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.047891 kubelet[3169]: W0813 00:28:54.047850 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.047891 kubelet[3169]: E0813 00:28:54.047856 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.047944 kubelet[3169]: E0813 00:28:54.047931 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.047944 kubelet[3169]: W0813 00:28:54.047935 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.047944 kubelet[3169]: E0813 00:28:54.047941 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.048098 kubelet[3169]: E0813 00:28:54.048085 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.048098 kubelet[3169]: W0813 00:28:54.048095 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.048171 kubelet[3169]: E0813 00:28:54.048103 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.048236 kubelet[3169]: E0813 00:28:54.048229 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.048266 kubelet[3169]: W0813 00:28:54.048242 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.048266 kubelet[3169]: E0813 00:28:54.048249 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.048357 kubelet[3169]: E0813 00:28:54.048349 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.048357 kubelet[3169]: W0813 00:28:54.048355 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.048421 kubelet[3169]: E0813 00:28:54.048361 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.048470 kubelet[3169]: E0813 00:28:54.048450 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.048498 kubelet[3169]: W0813 00:28:54.048470 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.048498 kubelet[3169]: E0813 00:28:54.048477 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.048641 kubelet[3169]: E0813 00:28:54.048632 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.048641 kubelet[3169]: W0813 00:28:54.048639 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.048739 kubelet[3169]: E0813 00:28:54.048646 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.048763 kubelet[3169]: E0813 00:28:54.048743 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.048763 kubelet[3169]: W0813 00:28:54.048747 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.048763 kubelet[3169]: E0813 00:28:54.048753 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.048858 kubelet[3169]: E0813 00:28:54.048851 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.048880 kubelet[3169]: W0813 00:28:54.048858 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.048880 kubelet[3169]: E0813 00:28:54.048864 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.048945 kubelet[3169]: E0813 00:28:54.048936 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.048945 kubelet[3169]: W0813 00:28:54.048940 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.049010 kubelet[3169]: E0813 00:28:54.048946 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.049063 kubelet[3169]: E0813 00:28:54.049054 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.049063 kubelet[3169]: W0813 00:28:54.049061 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.049113 kubelet[3169]: E0813 00:28:54.049068 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.049174 kubelet[3169]: E0813 00:28:54.049166 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.049174 kubelet[3169]: W0813 00:28:54.049173 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.049225 kubelet[3169]: E0813 00:28:54.049179 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.049277 kubelet[3169]: E0813 00:28:54.049269 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.049277 kubelet[3169]: W0813 00:28:54.049275 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.049357 kubelet[3169]: E0813 00:28:54.049281 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.049396 kubelet[3169]: E0813 00:28:54.049388 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.049396 kubelet[3169]: W0813 00:28:54.049394 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.049454 kubelet[3169]: E0813 00:28:54.049400 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.049491 kubelet[3169]: E0813 00:28:54.049476 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.049491 kubelet[3169]: W0813 00:28:54.049481 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.049491 kubelet[3169]: E0813 00:28:54.049487 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.049604 kubelet[3169]: E0813 00:28:54.049595 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.049604 kubelet[3169]: W0813 00:28:54.049601 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.049646 kubelet[3169]: E0813 00:28:54.049607 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.049913 kubelet[3169]: E0813 00:28:54.049857 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.049913 kubelet[3169]: W0813 00:28:54.049866 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.049913 kubelet[3169]: E0813 00:28:54.049875 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.050158 kubelet[3169]: E0813 00:28:54.050118 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.050158 kubelet[3169]: W0813 00:28:54.050127 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.050158 kubelet[3169]: E0813 00:28:54.050137 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.050450 kubelet[3169]: E0813 00:28:54.050406 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.050450 kubelet[3169]: W0813 00:28:54.050415 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.050450 kubelet[3169]: E0813 00:28:54.050425 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.050653 kubelet[3169]: E0813 00:28:54.050643 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.050653 kubelet[3169]: W0813 00:28:54.050652 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.050744 kubelet[3169]: E0813 00:28:54.050661 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.050871 kubelet[3169]: E0813 00:28:54.050859 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.050922 kubelet[3169]: W0813 00:28:54.050872 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.050922 kubelet[3169]: E0813 00:28:54.050880 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:54.055112 kubelet[3169]: E0813 00:28:54.055095 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:54.055112 kubelet[3169]: W0813 00:28:54.055108 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:54.055191 kubelet[3169]: E0813 00:28:54.055118 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:55.004377 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1329976303.mount: Deactivated successfully. Aug 13 00:28:55.905627 containerd[1720]: time="2025-08-13T00:28:55.905588024Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:55.907806 containerd[1720]: time="2025-08-13T00:28:55.907778530Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.2: active requests=0, bytes read=35233364" Aug 13 00:28:55.911204 containerd[1720]: time="2025-08-13T00:28:55.911167975Z" level=info msg="ImageCreate event name:\"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:55.915489 containerd[1720]: time="2025-08-13T00:28:55.914993015Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:55.915489 containerd[1720]: time="2025-08-13T00:28:55.915401799Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.2\" with image id \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\", size \"35233218\" in 2.274944893s" Aug 13 00:28:55.915489 containerd[1720]: time="2025-08-13T00:28:55.915425958Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\" returns image reference \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\"" Aug 13 00:28:55.916253 containerd[1720]: time="2025-08-13T00:28:55.916231824Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\"" Aug 13 00:28:55.932281 containerd[1720]: time="2025-08-13T00:28:55.932256389Z" level=info msg="CreateContainer within sandbox \"5762273d447720091720b906213ea2a5cc0ea18deeda7bf3430f6e59991700dc\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Aug 13 00:28:55.949485 containerd[1720]: time="2025-08-13T00:28:55.948396608Z" level=info msg="Container 2dacc505d176c15b58a6ea0199cc5a0d145a48da71a637a6df7fb2e56f1b7e8c: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:28:55.964627 containerd[1720]: time="2025-08-13T00:28:55.964594828Z" level=info msg="CreateContainer within sandbox \"5762273d447720091720b906213ea2a5cc0ea18deeda7bf3430f6e59991700dc\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"2dacc505d176c15b58a6ea0199cc5a0d145a48da71a637a6df7fb2e56f1b7e8c\"" Aug 13 00:28:55.965395 containerd[1720]: time="2025-08-13T00:28:55.965376948Z" level=info msg="StartContainer for \"2dacc505d176c15b58a6ea0199cc5a0d145a48da71a637a6df7fb2e56f1b7e8c\"" Aug 13 00:28:55.966285 containerd[1720]: time="2025-08-13T00:28:55.966261823Z" level=info msg="connecting to shim 2dacc505d176c15b58a6ea0199cc5a0d145a48da71a637a6df7fb2e56f1b7e8c" address="unix:///run/containerd/s/b274f104e21fb218cc08ec6ed14d28470ae715e2f5a792443394c4976b919c5d" protocol=ttrpc version=3 Aug 13 00:28:55.985430 systemd[1]: Started cri-containerd-2dacc505d176c15b58a6ea0199cc5a0d145a48da71a637a6df7fb2e56f1b7e8c.scope - libcontainer container 2dacc505d176c15b58a6ea0199cc5a0d145a48da71a637a6df7fb2e56f1b7e8c. Aug 13 00:28:56.022966 kubelet[3169]: E0813 00:28:56.022553 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-ptpbv" podUID="a62490c7-241c-47fd-bc6c-9b6d06fb91b7" Aug 13 00:28:56.029441 containerd[1720]: time="2025-08-13T00:28:56.029419532Z" level=info msg="StartContainer for \"2dacc505d176c15b58a6ea0199cc5a0d145a48da71a637a6df7fb2e56f1b7e8c\" returns successfully" Aug 13 00:28:56.153214 kubelet[3169]: E0813 00:28:56.153187 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.153214 kubelet[3169]: W0813 00:28:56.153208 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.153452 kubelet[3169]: E0813 00:28:56.153226 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.153452 kubelet[3169]: E0813 00:28:56.153349 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.153452 kubelet[3169]: W0813 00:28:56.153354 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.153452 kubelet[3169]: E0813 00:28:56.153362 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.153930 kubelet[3169]: E0813 00:28:56.153850 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.153930 kubelet[3169]: W0813 00:28:56.153861 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.153930 kubelet[3169]: E0813 00:28:56.153875 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.154403 kubelet[3169]: E0813 00:28:56.154380 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.154403 kubelet[3169]: W0813 00:28:56.154395 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.154484 kubelet[3169]: E0813 00:28:56.154407 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.154851 kubelet[3169]: E0813 00:28:56.154837 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.154913 kubelet[3169]: W0813 00:28:56.154851 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.154913 kubelet[3169]: E0813 00:28:56.154864 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.154999 kubelet[3169]: E0813 00:28:56.154968 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.154999 kubelet[3169]: W0813 00:28:56.154973 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.154999 kubelet[3169]: E0813 00:28:56.154980 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.156390 kubelet[3169]: E0813 00:28:56.156324 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.156390 kubelet[3169]: W0813 00:28:56.156339 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.156390 kubelet[3169]: E0813 00:28:56.156351 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.157106 kubelet[3169]: E0813 00:28:56.157090 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.157164 kubelet[3169]: W0813 00:28:56.157107 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.157164 kubelet[3169]: E0813 00:28:56.157119 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.157304 kubelet[3169]: E0813 00:28:56.157256 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.157304 kubelet[3169]: W0813 00:28:56.157264 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.157304 kubelet[3169]: E0813 00:28:56.157272 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.157419 kubelet[3169]: E0813 00:28:56.157377 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.157419 kubelet[3169]: W0813 00:28:56.157382 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.157419 kubelet[3169]: E0813 00:28:56.157388 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.157525 kubelet[3169]: E0813 00:28:56.157471 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.157525 kubelet[3169]: W0813 00:28:56.157476 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.157525 kubelet[3169]: E0813 00:28:56.157483 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.157680 kubelet[3169]: E0813 00:28:56.157578 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.157680 kubelet[3169]: W0813 00:28:56.157583 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.157680 kubelet[3169]: E0813 00:28:56.157588 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.157680 kubelet[3169]: E0813 00:28:56.157675 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.157680 kubelet[3169]: W0813 00:28:56.157679 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.157778 kubelet[3169]: E0813 00:28:56.157685 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.157778 kubelet[3169]: E0813 00:28:56.157759 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.157778 kubelet[3169]: W0813 00:28:56.157763 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.157778 kubelet[3169]: E0813 00:28:56.157768 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.157854 kubelet[3169]: E0813 00:28:56.157838 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.157854 kubelet[3169]: W0813 00:28:56.157842 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.157854 kubelet[3169]: E0813 00:28:56.157847 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.162173 kubelet[3169]: E0813 00:28:56.162158 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.162173 kubelet[3169]: W0813 00:28:56.162169 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.162339 kubelet[3169]: E0813 00:28:56.162180 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.162377 kubelet[3169]: E0813 00:28:56.162351 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.162377 kubelet[3169]: W0813 00:28:56.162358 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.162377 kubelet[3169]: E0813 00:28:56.162368 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.162608 kubelet[3169]: E0813 00:28:56.162493 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.162608 kubelet[3169]: W0813 00:28:56.162500 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.162608 kubelet[3169]: E0813 00:28:56.162507 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.162730 kubelet[3169]: E0813 00:28:56.162714 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.162730 kubelet[3169]: W0813 00:28:56.162721 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.162776 kubelet[3169]: E0813 00:28:56.162730 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.162987 kubelet[3169]: E0813 00:28:56.162923 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.162987 kubelet[3169]: W0813 00:28:56.162932 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.162987 kubelet[3169]: E0813 00:28:56.162941 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.163089 kubelet[3169]: E0813 00:28:56.163083 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.163127 kubelet[3169]: W0813 00:28:56.163122 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.163161 kubelet[3169]: E0813 00:28:56.163155 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.163378 kubelet[3169]: E0813 00:28:56.163353 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.163378 kubelet[3169]: W0813 00:28:56.163361 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.163378 kubelet[3169]: E0813 00:28:56.163369 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.163664 kubelet[3169]: E0813 00:28:56.163635 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.163664 kubelet[3169]: W0813 00:28:56.163645 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.163664 kubelet[3169]: E0813 00:28:56.163654 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.163775 kubelet[3169]: E0813 00:28:56.163767 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.163775 kubelet[3169]: W0813 00:28:56.163772 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.163841 kubelet[3169]: E0813 00:28:56.163778 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.163863 kubelet[3169]: E0813 00:28:56.163856 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.163863 kubelet[3169]: W0813 00:28:56.163860 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.163909 kubelet[3169]: E0813 00:28:56.163866 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.163986 kubelet[3169]: E0813 00:28:56.163976 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.163986 kubelet[3169]: W0813 00:28:56.163983 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.164031 kubelet[3169]: E0813 00:28:56.163995 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.164105 kubelet[3169]: E0813 00:28:56.164094 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.164105 kubelet[3169]: W0813 00:28:56.164102 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.164168 kubelet[3169]: E0813 00:28:56.164109 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.164503 kubelet[3169]: E0813 00:28:56.164252 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.164503 kubelet[3169]: W0813 00:28:56.164259 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.164503 kubelet[3169]: E0813 00:28:56.164266 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.164608 kubelet[3169]: E0813 00:28:56.164530 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.164608 kubelet[3169]: W0813 00:28:56.164537 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.164608 kubelet[3169]: E0813 00:28:56.164545 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.164671 kubelet[3169]: E0813 00:28:56.164644 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.164671 kubelet[3169]: W0813 00:28:56.164649 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.164671 kubelet[3169]: E0813 00:28:56.164655 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.164737 kubelet[3169]: E0813 00:28:56.164732 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.164760 kubelet[3169]: W0813 00:28:56.164737 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.164760 kubelet[3169]: E0813 00:28:56.164743 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.164846 kubelet[3169]: E0813 00:28:56.164837 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.164866 kubelet[3169]: W0813 00:28:56.164846 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.164866 kubelet[3169]: E0813 00:28:56.164852 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:56.165235 kubelet[3169]: E0813 00:28:56.165212 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:56.165235 kubelet[3169]: W0813 00:28:56.165229 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:56.165327 kubelet[3169]: E0813 00:28:56.165238 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.093479 kubelet[3169]: I0813 00:28:57.093452 3169 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:28:57.164711 kubelet[3169]: E0813 00:28:57.164614 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.164711 kubelet[3169]: W0813 00:28:57.164692 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.164711 kubelet[3169]: E0813 00:28:57.164709 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.164913 kubelet[3169]: E0813 00:28:57.164847 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.164913 kubelet[3169]: W0813 00:28:57.164852 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.164913 kubelet[3169]: E0813 00:28:57.164859 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.165082 kubelet[3169]: E0813 00:28:57.165045 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.165082 kubelet[3169]: W0813 00:28:57.165051 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.165082 kubelet[3169]: E0813 00:28:57.165058 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.165178 kubelet[3169]: E0813 00:28:57.165159 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.165178 kubelet[3169]: W0813 00:28:57.165164 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.165178 kubelet[3169]: E0813 00:28:57.165169 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.165337 kubelet[3169]: E0813 00:28:57.165321 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.165337 kubelet[3169]: W0813 00:28:57.165331 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.165400 kubelet[3169]: E0813 00:28:57.165340 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.165464 kubelet[3169]: E0813 00:28:57.165442 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.165464 kubelet[3169]: W0813 00:28:57.165462 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.165508 kubelet[3169]: E0813 00:28:57.165469 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.165578 kubelet[3169]: E0813 00:28:57.165558 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.165578 kubelet[3169]: W0813 00:28:57.165576 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.165621 kubelet[3169]: E0813 00:28:57.165581 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.165672 kubelet[3169]: E0813 00:28:57.165650 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.165672 kubelet[3169]: W0813 00:28:57.165670 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.165720 kubelet[3169]: E0813 00:28:57.165676 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.165786 kubelet[3169]: E0813 00:28:57.165770 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.165786 kubelet[3169]: W0813 00:28:57.165784 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.165828 kubelet[3169]: E0813 00:28:57.165790 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.165878 kubelet[3169]: E0813 00:28:57.165855 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.165878 kubelet[3169]: W0813 00:28:57.165875 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.165927 kubelet[3169]: E0813 00:28:57.165881 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.165961 kubelet[3169]: E0813 00:28:57.165953 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.165961 kubelet[3169]: W0813 00:28:57.165959 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.166008 kubelet[3169]: E0813 00:28:57.165964 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.166076 kubelet[3169]: E0813 00:28:57.166067 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.166076 kubelet[3169]: W0813 00:28:57.166075 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.166118 kubelet[3169]: E0813 00:28:57.166080 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.166162 kubelet[3169]: E0813 00:28:57.166152 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.166162 kubelet[3169]: W0813 00:28:57.166159 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.166207 kubelet[3169]: E0813 00:28:57.166165 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.166277 kubelet[3169]: E0813 00:28:57.166234 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.166277 kubelet[3169]: W0813 00:28:57.166238 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.166277 kubelet[3169]: E0813 00:28:57.166243 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.167461 kubelet[3169]: E0813 00:28:57.167443 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.167461 kubelet[3169]: W0813 00:28:57.167456 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.167608 kubelet[3169]: E0813 00:28:57.167469 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.169768 kubelet[3169]: E0813 00:28:57.169752 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.169768 kubelet[3169]: W0813 00:28:57.169763 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.169902 kubelet[3169]: E0813 00:28:57.169773 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.169936 kubelet[3169]: E0813 00:28:57.169919 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.169936 kubelet[3169]: W0813 00:28:57.169924 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.169936 kubelet[3169]: E0813 00:28:57.169931 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.170096 kubelet[3169]: E0813 00:28:57.170070 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.170096 kubelet[3169]: W0813 00:28:57.170092 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.170154 kubelet[3169]: E0813 00:28:57.170100 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.170215 kubelet[3169]: E0813 00:28:57.170192 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.170215 kubelet[3169]: W0813 00:28:57.170213 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.170271 kubelet[3169]: E0813 00:28:57.170219 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.170419 kubelet[3169]: E0813 00:28:57.170395 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.170419 kubelet[3169]: W0813 00:28:57.170417 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.170469 kubelet[3169]: E0813 00:28:57.170425 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.170565 kubelet[3169]: E0813 00:28:57.170555 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.170565 kubelet[3169]: W0813 00:28:57.170563 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.170611 kubelet[3169]: E0813 00:28:57.170570 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.170799 kubelet[3169]: E0813 00:28:57.170737 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.170799 kubelet[3169]: W0813 00:28:57.170747 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.170799 kubelet[3169]: E0813 00:28:57.170755 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.170876 kubelet[3169]: E0813 00:28:57.170845 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.170876 kubelet[3169]: W0813 00:28:57.170851 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.170876 kubelet[3169]: E0813 00:28:57.170857 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.170966 kubelet[3169]: E0813 00:28:57.170951 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.170966 kubelet[3169]: W0813 00:28:57.170962 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.171027 kubelet[3169]: E0813 00:28:57.170970 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.171079 kubelet[3169]: E0813 00:28:57.171048 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.171079 kubelet[3169]: W0813 00:28:57.171052 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.171079 kubelet[3169]: E0813 00:28:57.171058 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.171166 kubelet[3169]: E0813 00:28:57.171142 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.171166 kubelet[3169]: W0813 00:28:57.171147 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.171166 kubelet[3169]: E0813 00:28:57.171152 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.171265 kubelet[3169]: E0813 00:28:57.171225 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.171265 kubelet[3169]: W0813 00:28:57.171230 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.171265 kubelet[3169]: E0813 00:28:57.171235 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.171405 kubelet[3169]: E0813 00:28:57.171323 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.171405 kubelet[3169]: W0813 00:28:57.171326 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.171405 kubelet[3169]: E0813 00:28:57.171331 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.171462 kubelet[3169]: E0813 00:28:57.171428 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.171462 kubelet[3169]: W0813 00:28:57.171433 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.171462 kubelet[3169]: E0813 00:28:57.171440 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.171559 kubelet[3169]: E0813 00:28:57.171545 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.171559 kubelet[3169]: W0813 00:28:57.171553 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.171599 kubelet[3169]: E0813 00:28:57.171559 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.171756 kubelet[3169]: E0813 00:28:57.171746 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.171756 kubelet[3169]: W0813 00:28:57.171753 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.171816 kubelet[3169]: E0813 00:28:57.171759 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.172005 kubelet[3169]: E0813 00:28:57.171972 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.172005 kubelet[3169]: W0813 00:28:57.171981 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.172005 kubelet[3169]: E0813 00:28:57.171990 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.172116 kubelet[3169]: E0813 00:28:57.172109 3169 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:28:57.172116 kubelet[3169]: W0813 00:28:57.172115 3169 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:28:57.172168 kubelet[3169]: E0813 00:28:57.172121 3169 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:28:57.312121 containerd[1720]: time="2025-08-13T00:28:57.312091770Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:57.314395 containerd[1720]: time="2025-08-13T00:28:57.314365137Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2: active requests=0, bytes read=4446956" Aug 13 00:28:57.317242 containerd[1720]: time="2025-08-13T00:28:57.317170435Z" level=info msg="ImageCreate event name:\"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:57.320427 containerd[1720]: time="2025-08-13T00:28:57.320401354Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:28:57.320772 containerd[1720]: time="2025-08-13T00:28:57.320687194Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" with image id \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\", size \"5939619\" in 1.404427316s" Aug 13 00:28:57.320772 containerd[1720]: time="2025-08-13T00:28:57.320710761Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" returns image reference \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\"" Aug 13 00:28:57.326796 containerd[1720]: time="2025-08-13T00:28:57.326771872Z" level=info msg="CreateContainer within sandbox \"c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Aug 13 00:28:57.356600 containerd[1720]: time="2025-08-13T00:28:57.356314403Z" level=info msg="Container 58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:28:57.372931 containerd[1720]: time="2025-08-13T00:28:57.372906126Z" level=info msg="CreateContainer within sandbox \"c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852\"" Aug 13 00:28:57.373321 containerd[1720]: time="2025-08-13T00:28:57.373274573Z" level=info msg="StartContainer for \"58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852\"" Aug 13 00:28:57.374678 containerd[1720]: time="2025-08-13T00:28:57.374637440Z" level=info msg="connecting to shim 58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852" address="unix:///run/containerd/s/5c173d3927e4e19eb263c99381ddee7ad8a6c5ea45d5176f2a82cc7983b8a868" protocol=ttrpc version=3 Aug 13 00:28:57.393403 systemd[1]: Started cri-containerd-58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852.scope - libcontainer container 58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852. Aug 13 00:28:57.425225 containerd[1720]: time="2025-08-13T00:28:57.425183220Z" level=info msg="StartContainer for \"58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852\" returns successfully" Aug 13 00:28:57.430968 systemd[1]: cri-containerd-58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852.scope: Deactivated successfully. Aug 13 00:28:57.433944 containerd[1720]: time="2025-08-13T00:28:57.433913723Z" level=info msg="TaskExit event in podsandbox handler container_id:\"58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852\" id:\"58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852\" pid:3880 exited_at:{seconds:1755044937 nanos:433586123}" Aug 13 00:28:57.434014 containerd[1720]: time="2025-08-13T00:28:57.433964385Z" level=info msg="received exit event container_id:\"58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852\" id:\"58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852\" pid:3880 exited_at:{seconds:1755044937 nanos:433586123}" Aug 13 00:28:57.449202 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-58df448136ff23d9e10100ce171be3a726c26719d192b2ee756103cc2f565852-rootfs.mount: Deactivated successfully. Aug 13 00:28:58.021821 kubelet[3169]: E0813 00:28:58.021783 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-ptpbv" podUID="a62490c7-241c-47fd-bc6c-9b6d06fb91b7" Aug 13 00:28:58.114116 kubelet[3169]: I0813 00:28:58.113962 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-658b76bf5b-9wx5r" podStartSLOduration=2.837763616 podStartE2EDuration="5.11394437s" podCreationTimestamp="2025-08-13 00:28:53 +0000 UTC" firstStartedPulling="2025-08-13 00:28:53.639931749 +0000 UTC m=+19.698962840" lastFinishedPulling="2025-08-13 00:28:55.916112502 +0000 UTC m=+21.975143594" observedRunningTime="2025-08-13 00:28:56.106813718 +0000 UTC m=+22.165844804" watchObservedRunningTime="2025-08-13 00:28:58.11394437 +0000 UTC m=+24.172975462" Aug 13 00:29:00.022594 kubelet[3169]: E0813 00:29:00.022306 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-ptpbv" podUID="a62490c7-241c-47fd-bc6c-9b6d06fb91b7" Aug 13 00:29:00.106215 containerd[1720]: time="2025-08-13T00:29:00.106180290Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\"" Aug 13 00:29:02.023865 kubelet[3169]: E0813 00:29:02.023832 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-ptpbv" podUID="a62490c7-241c-47fd-bc6c-9b6d06fb91b7" Aug 13 00:29:02.611250 kubelet[3169]: I0813 00:29:02.611173 3169 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:29:04.024328 kubelet[3169]: E0813 00:29:04.024266 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-ptpbv" podUID="a62490c7-241c-47fd-bc6c-9b6d06fb91b7" Aug 13 00:29:04.144298 containerd[1720]: time="2025-08-13T00:29:04.144259297Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:04.147157 containerd[1720]: time="2025-08-13T00:29:04.147128419Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.2: active requests=0, bytes read=70436221" Aug 13 00:29:04.150153 containerd[1720]: time="2025-08-13T00:29:04.150100215Z" level=info msg="ImageCreate event name:\"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:04.153817 containerd[1720]: time="2025-08-13T00:29:04.153751781Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:04.154214 containerd[1720]: time="2025-08-13T00:29:04.154060360Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.2\" with image id \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\", size \"71928924\" in 4.047841377s" Aug 13 00:29:04.154214 containerd[1720]: time="2025-08-13T00:29:04.154084836Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\" returns image reference \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\"" Aug 13 00:29:04.160383 containerd[1720]: time="2025-08-13T00:29:04.160353443Z" level=info msg="CreateContainer within sandbox \"c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Aug 13 00:29:04.180307 containerd[1720]: time="2025-08-13T00:29:04.179870510Z" level=info msg="Container 864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:04.197448 containerd[1720]: time="2025-08-13T00:29:04.197424523Z" level=info msg="CreateContainer within sandbox \"c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0\"" Aug 13 00:29:04.197821 containerd[1720]: time="2025-08-13T00:29:04.197774991Z" level=info msg="StartContainer for \"864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0\"" Aug 13 00:29:04.199215 containerd[1720]: time="2025-08-13T00:29:04.199184664Z" level=info msg="connecting to shim 864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0" address="unix:///run/containerd/s/5c173d3927e4e19eb263c99381ddee7ad8a6c5ea45d5176f2a82cc7983b8a868" protocol=ttrpc version=3 Aug 13 00:29:04.219406 systemd[1]: Started cri-containerd-864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0.scope - libcontainer container 864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0. Aug 13 00:29:04.252710 containerd[1720]: time="2025-08-13T00:29:04.251840437Z" level=info msg="StartContainer for \"864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0\" returns successfully" Aug 13 00:29:05.376941 containerd[1720]: time="2025-08-13T00:29:05.376893489Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 00:29:05.378502 systemd[1]: cri-containerd-864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0.scope: Deactivated successfully. Aug 13 00:29:05.378736 systemd[1]: cri-containerd-864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0.scope: Consumed 364ms CPU time, 192.7M memory peak, 171.2M written to disk. Aug 13 00:29:05.380929 containerd[1720]: time="2025-08-13T00:29:05.380870212Z" level=info msg="received exit event container_id:\"864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0\" id:\"864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0\" pid:3943 exited_at:{seconds:1755044945 nanos:380670340}" Aug 13 00:29:05.381105 containerd[1720]: time="2025-08-13T00:29:05.381011343Z" level=info msg="TaskExit event in podsandbox handler container_id:\"864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0\" id:\"864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0\" pid:3943 exited_at:{seconds:1755044945 nanos:380670340}" Aug 13 00:29:05.396904 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-864e893a95f3e4412df6b4a5155230b89b0d2c24334545c30e7db53ba94fd8d0-rootfs.mount: Deactivated successfully. Aug 13 00:29:05.465656 kubelet[3169]: I0813 00:29:05.465637 3169 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Aug 13 00:29:05.649184 systemd[1]: Created slice kubepods-burstable-pod2680547d_4498_4eac_9daa_64c89f065d86.slice - libcontainer container kubepods-burstable-pod2680547d_4498_4eac_9daa_64c89f065d86.slice. Aug 13 00:29:05.723374 kubelet[3169]: I0813 00:29:05.723330 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcdgp\" (UniqueName: \"kubernetes.io/projected/2680547d-4498-4eac-9daa-64c89f065d86-kube-api-access-vcdgp\") pod \"coredns-674b8bbfcf-hhdbb\" (UID: \"2680547d-4498-4eac-9daa-64c89f065d86\") " pod="kube-system/coredns-674b8bbfcf-hhdbb" Aug 13 00:29:05.723374 kubelet[3169]: I0813 00:29:05.723360 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2680547d-4498-4eac-9daa-64c89f065d86-config-volume\") pod \"coredns-674b8bbfcf-hhdbb\" (UID: \"2680547d-4498-4eac-9daa-64c89f065d86\") " pod="kube-system/coredns-674b8bbfcf-hhdbb" Aug 13 00:29:05.952266 containerd[1720]: time="2025-08-13T00:29:05.952033020Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hhdbb,Uid:2680547d-4498-4eac-9daa-64c89f065d86,Namespace:kube-system,Attempt:0,}" Aug 13 00:29:05.957578 systemd[1]: Created slice kubepods-burstable-pod71824225_ae9f_4348_bcd0_e9d39f3bdee6.slice - libcontainer container kubepods-burstable-pod71824225_ae9f_4348_bcd0_e9d39f3bdee6.slice. Aug 13 00:29:06.024505 kubelet[3169]: I0813 00:29:06.024478 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn9mv\" (UniqueName: \"kubernetes.io/projected/71824225-ae9f-4348-bcd0-e9d39f3bdee6-kube-api-access-rn9mv\") pod \"coredns-674b8bbfcf-nbpw8\" (UID: \"71824225-ae9f-4348-bcd0-e9d39f3bdee6\") " pod="kube-system/coredns-674b8bbfcf-nbpw8" Aug 13 00:29:06.024505 kubelet[3169]: I0813 00:29:06.024509 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71824225-ae9f-4348-bcd0-e9d39f3bdee6-config-volume\") pod \"coredns-674b8bbfcf-nbpw8\" (UID: \"71824225-ae9f-4348-bcd0-e9d39f3bdee6\") " pod="kube-system/coredns-674b8bbfcf-nbpw8" Aug 13 00:29:06.264464 containerd[1720]: time="2025-08-13T00:29:06.263940782Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-nbpw8,Uid:71824225-ae9f-4348-bcd0-e9d39f3bdee6,Namespace:kube-system,Attempt:0,}" Aug 13 00:29:06.264283 systemd[1]: Created slice kubepods-besteffort-pode13962cc_6aa3_411a_81be_4a870f09c685.slice - libcontainer container kubepods-besteffort-pode13962cc_6aa3_411a_81be_4a870f09c685.slice. Aug 13 00:29:06.294272 systemd[1]: Created slice kubepods-besteffort-pod3243471f_918a_41b7_bcb5_e0d349b40c90.slice - libcontainer container kubepods-besteffort-pod3243471f_918a_41b7_bcb5_e0d349b40c90.slice. Aug 13 00:29:06.304042 systemd[1]: Created slice kubepods-besteffort-podeaaa778d_497a_49e1_b653_c6dedc50adee.slice - libcontainer container kubepods-besteffort-podeaaa778d_497a_49e1_b653_c6dedc50adee.slice. Aug 13 00:29:06.312762 systemd[1]: Created slice kubepods-besteffort-poda62490c7_241c_47fd_bc6c_9b6d06fb91b7.slice - libcontainer container kubepods-besteffort-poda62490c7_241c_47fd_bc6c_9b6d06fb91b7.slice. Aug 13 00:29:06.320821 containerd[1720]: time="2025-08-13T00:29:06.320679336Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-ptpbv,Uid:a62490c7-241c-47fd-bc6c-9b6d06fb91b7,Namespace:calico-system,Attempt:0,}" Aug 13 00:29:06.328016 systemd[1]: Created slice kubepods-besteffort-podf36fa87a_1607_484a_99a0_3c120173b24c.slice - libcontainer container kubepods-besteffort-podf36fa87a_1607_484a_99a0_3c120173b24c.slice. Aug 13 00:29:06.329703 kubelet[3169]: I0813 00:29:06.329461 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-whisker-backend-key-pair\") pod \"whisker-64f59c8569-pqkz6\" (UID: \"f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc\") " pod="calico-system/whisker-64f59c8569-pqkz6" Aug 13 00:29:06.329703 kubelet[3169]: I0813 00:29:06.329497 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaaa778d-497a-49e1-b653-c6dedc50adee-config\") pod \"goldmane-768f4c5c69-4c59z\" (UID: \"eaaa778d-497a-49e1-b653-c6dedc50adee\") " pod="calico-system/goldmane-768f4c5c69-4c59z" Aug 13 00:29:06.329703 kubelet[3169]: I0813 00:29:06.329536 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnjzq\" (UniqueName: \"kubernetes.io/projected/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-kube-api-access-cnjzq\") pod \"whisker-64f59c8569-pqkz6\" (UID: \"f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc\") " pod="calico-system/whisker-64f59c8569-pqkz6" Aug 13 00:29:06.329703 kubelet[3169]: I0813 00:29:06.329556 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/f36fa87a-1607-484a-99a0-3c120173b24c-calico-apiserver-certs\") pod \"calico-apiserver-7d685458f5-h9nsv\" (UID: \"f36fa87a-1607-484a-99a0-3c120173b24c\") " pod="calico-apiserver/calico-apiserver-7d685458f5-h9nsv" Aug 13 00:29:06.329703 kubelet[3169]: I0813 00:29:06.329577 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhlnk\" (UniqueName: \"kubernetes.io/projected/f36fa87a-1607-484a-99a0-3c120173b24c-kube-api-access-nhlnk\") pod \"calico-apiserver-7d685458f5-h9nsv\" (UID: \"f36fa87a-1607-484a-99a0-3c120173b24c\") " pod="calico-apiserver/calico-apiserver-7d685458f5-h9nsv" Aug 13 00:29:06.329871 kubelet[3169]: I0813 00:29:06.329598 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-whisker-ca-bundle\") pod \"whisker-64f59c8569-pqkz6\" (UID: \"f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc\") " pod="calico-system/whisker-64f59c8569-pqkz6" Aug 13 00:29:06.329871 kubelet[3169]: I0813 00:29:06.329619 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp7vg\" (UniqueName: \"kubernetes.io/projected/eaaa778d-497a-49e1-b653-c6dedc50adee-kube-api-access-hp7vg\") pod \"goldmane-768f4c5c69-4c59z\" (UID: \"eaaa778d-497a-49e1-b653-c6dedc50adee\") " pod="calico-system/goldmane-768f4c5c69-4c59z" Aug 13 00:29:06.329871 kubelet[3169]: I0813 00:29:06.329640 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e13962cc-6aa3-411a-81be-4a870f09c685-tigera-ca-bundle\") pod \"calico-kube-controllers-5fd87f6d8c-qtj5z\" (UID: \"e13962cc-6aa3-411a-81be-4a870f09c685\") " pod="calico-system/calico-kube-controllers-5fd87f6d8c-qtj5z" Aug 13 00:29:06.329871 kubelet[3169]: I0813 00:29:06.329658 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/3243471f-918a-41b7-bcb5-e0d349b40c90-calico-apiserver-certs\") pod \"calico-apiserver-7d685458f5-fchq5\" (UID: \"3243471f-918a-41b7-bcb5-e0d349b40c90\") " pod="calico-apiserver/calico-apiserver-7d685458f5-fchq5" Aug 13 00:29:06.329871 kubelet[3169]: I0813 00:29:06.329677 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7k2l\" (UniqueName: \"kubernetes.io/projected/e13962cc-6aa3-411a-81be-4a870f09c685-kube-api-access-h7k2l\") pod \"calico-kube-controllers-5fd87f6d8c-qtj5z\" (UID: \"e13962cc-6aa3-411a-81be-4a870f09c685\") " pod="calico-system/calico-kube-controllers-5fd87f6d8c-qtj5z" Aug 13 00:29:06.329986 kubelet[3169]: I0813 00:29:06.329696 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4rqq\" (UniqueName: \"kubernetes.io/projected/3243471f-918a-41b7-bcb5-e0d349b40c90-kube-api-access-m4rqq\") pod \"calico-apiserver-7d685458f5-fchq5\" (UID: \"3243471f-918a-41b7-bcb5-e0d349b40c90\") " pod="calico-apiserver/calico-apiserver-7d685458f5-fchq5" Aug 13 00:29:06.329986 kubelet[3169]: I0813 00:29:06.329716 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eaaa778d-497a-49e1-b653-c6dedc50adee-goldmane-ca-bundle\") pod \"goldmane-768f4c5c69-4c59z\" (UID: \"eaaa778d-497a-49e1-b653-c6dedc50adee\") " pod="calico-system/goldmane-768f4c5c69-4c59z" Aug 13 00:29:06.329986 kubelet[3169]: I0813 00:29:06.329737 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/eaaa778d-497a-49e1-b653-c6dedc50adee-goldmane-key-pair\") pod \"goldmane-768f4c5c69-4c59z\" (UID: \"eaaa778d-497a-49e1-b653-c6dedc50adee\") " pod="calico-system/goldmane-768f4c5c69-4c59z" Aug 13 00:29:06.337383 systemd[1]: Created slice kubepods-besteffort-podf1e6c903_9f98_4e5f_acbb_c0e2d3c613dc.slice - libcontainer container kubepods-besteffort-podf1e6c903_9f98_4e5f_acbb_c0e2d3c613dc.slice. Aug 13 00:29:06.367727 containerd[1720]: time="2025-08-13T00:29:06.367645472Z" level=error msg="Failed to destroy network for sandbox \"7399692a1133ec5996ecd4d10f7869dbefddb5e584a3721ea644d18c17f0cde8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.370629 containerd[1720]: time="2025-08-13T00:29:06.370591619Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hhdbb,Uid:2680547d-4498-4eac-9daa-64c89f065d86,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7399692a1133ec5996ecd4d10f7869dbefddb5e584a3721ea644d18c17f0cde8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.370800 kubelet[3169]: E0813 00:29:06.370766 3169 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7399692a1133ec5996ecd4d10f7869dbefddb5e584a3721ea644d18c17f0cde8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.370843 kubelet[3169]: E0813 00:29:06.370829 3169 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7399692a1133ec5996ecd4d10f7869dbefddb5e584a3721ea644d18c17f0cde8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-hhdbb" Aug 13 00:29:06.371124 kubelet[3169]: E0813 00:29:06.370846 3169 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7399692a1133ec5996ecd4d10f7869dbefddb5e584a3721ea644d18c17f0cde8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-hhdbb" Aug 13 00:29:06.371124 kubelet[3169]: E0813 00:29:06.370908 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-hhdbb_kube-system(2680547d-4498-4eac-9daa-64c89f065d86)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-hhdbb_kube-system(2680547d-4498-4eac-9daa-64c89f065d86)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7399692a1133ec5996ecd4d10f7869dbefddb5e584a3721ea644d18c17f0cde8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-hhdbb" podUID="2680547d-4498-4eac-9daa-64c89f065d86" Aug 13 00:29:06.382396 containerd[1720]: time="2025-08-13T00:29:06.382359405Z" level=error msg="Failed to destroy network for sandbox \"fc7a9c5262bd95293f23ebad90bfb4b5c1a9470e7701b80e6cee12c6c76999e2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.385945 containerd[1720]: time="2025-08-13T00:29:06.385913402Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-nbpw8,Uid:71824225-ae9f-4348-bcd0-e9d39f3bdee6,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"fc7a9c5262bd95293f23ebad90bfb4b5c1a9470e7701b80e6cee12c6c76999e2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.386195 kubelet[3169]: E0813 00:29:06.386165 3169 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fc7a9c5262bd95293f23ebad90bfb4b5c1a9470e7701b80e6cee12c6c76999e2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.386246 kubelet[3169]: E0813 00:29:06.386205 3169 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fc7a9c5262bd95293f23ebad90bfb4b5c1a9470e7701b80e6cee12c6c76999e2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-nbpw8" Aug 13 00:29:06.386246 kubelet[3169]: E0813 00:29:06.386225 3169 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fc7a9c5262bd95293f23ebad90bfb4b5c1a9470e7701b80e6cee12c6c76999e2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-nbpw8" Aug 13 00:29:06.386787 kubelet[3169]: E0813 00:29:06.386263 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-nbpw8_kube-system(71824225-ae9f-4348-bcd0-e9d39f3bdee6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-nbpw8_kube-system(71824225-ae9f-4348-bcd0-e9d39f3bdee6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fc7a9c5262bd95293f23ebad90bfb4b5c1a9470e7701b80e6cee12c6c76999e2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-nbpw8" podUID="71824225-ae9f-4348-bcd0-e9d39f3bdee6" Aug 13 00:29:06.396913 systemd[1]: run-netns-cni\x2d8bf0b6d3\x2df58d\x2d5b96\x2d9463\x2dc3a042944aeb.mount: Deactivated successfully. Aug 13 00:29:06.397255 systemd[1]: run-netns-cni\x2d3ae16810\x2d2c77\x2d4c4d\x2d8918\x2d8f7cfcc3a830.mount: Deactivated successfully. Aug 13 00:29:06.399021 containerd[1720]: time="2025-08-13T00:29:06.398989692Z" level=error msg="Failed to destroy network for sandbox \"1701f1af6192e6d105ed35c44b210e125d16873b693ffdbbf5a87e641147a0ce\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.400510 systemd[1]: run-netns-cni\x2d1f49f07d\x2ddd05\x2d8a9e\x2d680b\x2d1c2310bca905.mount: Deactivated successfully. Aug 13 00:29:06.403769 containerd[1720]: time="2025-08-13T00:29:06.403739663Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-ptpbv,Uid:a62490c7-241c-47fd-bc6c-9b6d06fb91b7,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"1701f1af6192e6d105ed35c44b210e125d16873b693ffdbbf5a87e641147a0ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.403903 kubelet[3169]: E0813 00:29:06.403880 3169 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1701f1af6192e6d105ed35c44b210e125d16873b693ffdbbf5a87e641147a0ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.403946 kubelet[3169]: E0813 00:29:06.403932 3169 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1701f1af6192e6d105ed35c44b210e125d16873b693ffdbbf5a87e641147a0ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-ptpbv" Aug 13 00:29:06.403985 kubelet[3169]: E0813 00:29:06.403952 3169 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1701f1af6192e6d105ed35c44b210e125d16873b693ffdbbf5a87e641147a0ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-ptpbv" Aug 13 00:29:06.404023 kubelet[3169]: E0813 00:29:06.404003 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-ptpbv_calico-system(a62490c7-241c-47fd-bc6c-9b6d06fb91b7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-ptpbv_calico-system(a62490c7-241c-47fd-bc6c-9b6d06fb91b7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1701f1af6192e6d105ed35c44b210e125d16873b693ffdbbf5a87e641147a0ce\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-ptpbv" podUID="a62490c7-241c-47fd-bc6c-9b6d06fb91b7" Aug 13 00:29:06.569871 containerd[1720]: time="2025-08-13T00:29:06.569801451Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5fd87f6d8c-qtj5z,Uid:e13962cc-6aa3-411a-81be-4a870f09c685,Namespace:calico-system,Attempt:0,}" Aug 13 00:29:06.600309 containerd[1720]: time="2025-08-13T00:29:06.600249728Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d685458f5-fchq5,Uid:3243471f-918a-41b7-bcb5-e0d349b40c90,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:29:06.606622 containerd[1720]: time="2025-08-13T00:29:06.606590955Z" level=error msg="Failed to destroy network for sandbox \"e25b031af92721001878623b268198a722ca09e805146014a8b79b80540bf5b8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.608992 containerd[1720]: time="2025-08-13T00:29:06.608960404Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-4c59z,Uid:eaaa778d-497a-49e1-b653-c6dedc50adee,Namespace:calico-system,Attempt:0,}" Aug 13 00:29:06.610317 containerd[1720]: time="2025-08-13T00:29:06.610225282Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5fd87f6d8c-qtj5z,Uid:e13962cc-6aa3-411a-81be-4a870f09c685,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e25b031af92721001878623b268198a722ca09e805146014a8b79b80540bf5b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.610847 kubelet[3169]: E0813 00:29:06.610612 3169 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e25b031af92721001878623b268198a722ca09e805146014a8b79b80540bf5b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.610847 kubelet[3169]: E0813 00:29:06.610661 3169 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e25b031af92721001878623b268198a722ca09e805146014a8b79b80540bf5b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5fd87f6d8c-qtj5z" Aug 13 00:29:06.610847 kubelet[3169]: E0813 00:29:06.610682 3169 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e25b031af92721001878623b268198a722ca09e805146014a8b79b80540bf5b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5fd87f6d8c-qtj5z" Aug 13 00:29:06.611166 kubelet[3169]: E0813 00:29:06.610732 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5fd87f6d8c-qtj5z_calico-system(e13962cc-6aa3-411a-81be-4a870f09c685)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5fd87f6d8c-qtj5z_calico-system(e13962cc-6aa3-411a-81be-4a870f09c685)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e25b031af92721001878623b268198a722ca09e805146014a8b79b80540bf5b8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5fd87f6d8c-qtj5z" podUID="e13962cc-6aa3-411a-81be-4a870f09c685" Aug 13 00:29:06.634587 containerd[1720]: time="2025-08-13T00:29:06.634546757Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d685458f5-h9nsv,Uid:f36fa87a-1607-484a-99a0-3c120173b24c,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:29:06.643485 containerd[1720]: time="2025-08-13T00:29:06.643427888Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-64f59c8569-pqkz6,Uid:f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc,Namespace:calico-system,Attempt:0,}" Aug 13 00:29:06.664916 containerd[1720]: time="2025-08-13T00:29:06.664884721Z" level=error msg="Failed to destroy network for sandbox \"783dc7bf2b6872c2b48154cbf3ffaf732c240545c0dc469a23aafcd2f6e7d53f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.667858 containerd[1720]: time="2025-08-13T00:29:06.667740847Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d685458f5-fchq5,Uid:3243471f-918a-41b7-bcb5-e0d349b40c90,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"783dc7bf2b6872c2b48154cbf3ffaf732c240545c0dc469a23aafcd2f6e7d53f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.668692 kubelet[3169]: E0813 00:29:06.668360 3169 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"783dc7bf2b6872c2b48154cbf3ffaf732c240545c0dc469a23aafcd2f6e7d53f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.668692 kubelet[3169]: E0813 00:29:06.668414 3169 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"783dc7bf2b6872c2b48154cbf3ffaf732c240545c0dc469a23aafcd2f6e7d53f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d685458f5-fchq5" Aug 13 00:29:06.668692 kubelet[3169]: E0813 00:29:06.668434 3169 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"783dc7bf2b6872c2b48154cbf3ffaf732c240545c0dc469a23aafcd2f6e7d53f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d685458f5-fchq5" Aug 13 00:29:06.668820 kubelet[3169]: E0813 00:29:06.668488 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d685458f5-fchq5_calico-apiserver(3243471f-918a-41b7-bcb5-e0d349b40c90)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d685458f5-fchq5_calico-apiserver(3243471f-918a-41b7-bcb5-e0d349b40c90)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"783dc7bf2b6872c2b48154cbf3ffaf732c240545c0dc469a23aafcd2f6e7d53f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d685458f5-fchq5" podUID="3243471f-918a-41b7-bcb5-e0d349b40c90" Aug 13 00:29:06.689215 containerd[1720]: time="2025-08-13T00:29:06.689188125Z" level=error msg="Failed to destroy network for sandbox \"535e87901fdc3fbaae8ffadbd0dd8306ffa3e9d30c4eac5724beb53d2ce83c8c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.692708 containerd[1720]: time="2025-08-13T00:29:06.692667763Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-4c59z,Uid:eaaa778d-497a-49e1-b653-c6dedc50adee,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"535e87901fdc3fbaae8ffadbd0dd8306ffa3e9d30c4eac5724beb53d2ce83c8c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.693007 kubelet[3169]: E0813 00:29:06.692974 3169 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"535e87901fdc3fbaae8ffadbd0dd8306ffa3e9d30c4eac5724beb53d2ce83c8c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.693062 kubelet[3169]: E0813 00:29:06.693024 3169 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"535e87901fdc3fbaae8ffadbd0dd8306ffa3e9d30c4eac5724beb53d2ce83c8c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-768f4c5c69-4c59z" Aug 13 00:29:06.693062 kubelet[3169]: E0813 00:29:06.693043 3169 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"535e87901fdc3fbaae8ffadbd0dd8306ffa3e9d30c4eac5724beb53d2ce83c8c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-768f4c5c69-4c59z" Aug 13 00:29:06.694227 kubelet[3169]: E0813 00:29:06.694155 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-768f4c5c69-4c59z_calico-system(eaaa778d-497a-49e1-b653-c6dedc50adee)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-768f4c5c69-4c59z_calico-system(eaaa778d-497a-49e1-b653-c6dedc50adee)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"535e87901fdc3fbaae8ffadbd0dd8306ffa3e9d30c4eac5724beb53d2ce83c8c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-768f4c5c69-4c59z" podUID="eaaa778d-497a-49e1-b653-c6dedc50adee" Aug 13 00:29:06.708806 containerd[1720]: time="2025-08-13T00:29:06.708715177Z" level=error msg="Failed to destroy network for sandbox \"ce9ba367fe56b4ce9efd4e6ae6226eec2c681d83fc89bd5e44de27edb8d5ddc3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.712545 containerd[1720]: time="2025-08-13T00:29:06.712474185Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d685458f5-h9nsv,Uid:f36fa87a-1607-484a-99a0-3c120173b24c,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ce9ba367fe56b4ce9efd4e6ae6226eec2c681d83fc89bd5e44de27edb8d5ddc3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.712732 kubelet[3169]: E0813 00:29:06.712705 3169 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ce9ba367fe56b4ce9efd4e6ae6226eec2c681d83fc89bd5e44de27edb8d5ddc3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.712817 kubelet[3169]: E0813 00:29:06.712806 3169 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ce9ba367fe56b4ce9efd4e6ae6226eec2c681d83fc89bd5e44de27edb8d5ddc3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d685458f5-h9nsv" Aug 13 00:29:06.712901 kubelet[3169]: E0813 00:29:06.712869 3169 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ce9ba367fe56b4ce9efd4e6ae6226eec2c681d83fc89bd5e44de27edb8d5ddc3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d685458f5-h9nsv" Aug 13 00:29:06.713010 kubelet[3169]: E0813 00:29:06.712993 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d685458f5-h9nsv_calico-apiserver(f36fa87a-1607-484a-99a0-3c120173b24c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d685458f5-h9nsv_calico-apiserver(f36fa87a-1607-484a-99a0-3c120173b24c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ce9ba367fe56b4ce9efd4e6ae6226eec2c681d83fc89bd5e44de27edb8d5ddc3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d685458f5-h9nsv" podUID="f36fa87a-1607-484a-99a0-3c120173b24c" Aug 13 00:29:06.713871 containerd[1720]: time="2025-08-13T00:29:06.713834567Z" level=error msg="Failed to destroy network for sandbox \"b2d636ff2a495acf8f82a0c812f682c99a23a2d0fa9eb71df9b535617cceb91b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.726262 containerd[1720]: time="2025-08-13T00:29:06.726232735Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-64f59c8569-pqkz6,Uid:f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b2d636ff2a495acf8f82a0c812f682c99a23a2d0fa9eb71df9b535617cceb91b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.726402 kubelet[3169]: E0813 00:29:06.726373 3169 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b2d636ff2a495acf8f82a0c812f682c99a23a2d0fa9eb71df9b535617cceb91b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:29:06.726440 kubelet[3169]: E0813 00:29:06.726414 3169 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b2d636ff2a495acf8f82a0c812f682c99a23a2d0fa9eb71df9b535617cceb91b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-64f59c8569-pqkz6" Aug 13 00:29:06.726440 kubelet[3169]: E0813 00:29:06.726430 3169 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b2d636ff2a495acf8f82a0c812f682c99a23a2d0fa9eb71df9b535617cceb91b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-64f59c8569-pqkz6" Aug 13 00:29:06.726494 kubelet[3169]: E0813 00:29:06.726463 3169 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-64f59c8569-pqkz6_calico-system(f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-64f59c8569-pqkz6_calico-system(f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b2d636ff2a495acf8f82a0c812f682c99a23a2d0fa9eb71df9b535617cceb91b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-64f59c8569-pqkz6" podUID="f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc" Aug 13 00:29:07.122248 containerd[1720]: time="2025-08-13T00:29:07.122214254Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\"" Aug 13 00:29:14.095848 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2220831134.mount: Deactivated successfully. Aug 13 00:29:14.137662 containerd[1720]: time="2025-08-13T00:29:14.137627066Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:14.139989 containerd[1720]: time="2025-08-13T00:29:14.139952501Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.2: active requests=0, bytes read=158500163" Aug 13 00:29:14.142584 containerd[1720]: time="2025-08-13T00:29:14.142545818Z" level=info msg="ImageCreate event name:\"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:14.145740 containerd[1720]: time="2025-08-13T00:29:14.145702331Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:14.146074 containerd[1720]: time="2025-08-13T00:29:14.145980046Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.2\" with image id \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\", size \"158500025\" in 7.023723618s" Aug 13 00:29:14.146074 containerd[1720]: time="2025-08-13T00:29:14.146009188Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\" returns image reference \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\"" Aug 13 00:29:14.160478 containerd[1720]: time="2025-08-13T00:29:14.160451110Z" level=info msg="CreateContainer within sandbox \"c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Aug 13 00:29:14.183158 containerd[1720]: time="2025-08-13T00:29:14.181614024Z" level=info msg="Container 3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:14.186978 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount254764595.mount: Deactivated successfully. Aug 13 00:29:14.208308 containerd[1720]: time="2025-08-13T00:29:14.208273232Z" level=info msg="CreateContainer within sandbox \"c69254e1718ab70a1d3a8454b62a10cf7c9661818cb13ecde288d113a5b2f6e4\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3\"" Aug 13 00:29:14.208789 containerd[1720]: time="2025-08-13T00:29:14.208698197Z" level=info msg="StartContainer for \"3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3\"" Aug 13 00:29:14.212411 containerd[1720]: time="2025-08-13T00:29:14.212377116Z" level=info msg="connecting to shim 3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3" address="unix:///run/containerd/s/5c173d3927e4e19eb263c99381ddee7ad8a6c5ea45d5176f2a82cc7983b8a868" protocol=ttrpc version=3 Aug 13 00:29:14.232425 systemd[1]: Started cri-containerd-3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3.scope - libcontainer container 3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3. Aug 13 00:29:14.262586 containerd[1720]: time="2025-08-13T00:29:14.262561044Z" level=info msg="StartContainer for \"3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3\" returns successfully" Aug 13 00:29:14.763421 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Aug 13 00:29:14.763502 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Aug 13 00:29:14.973618 kubelet[3169]: I0813 00:29:14.973575 3169 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnjzq\" (UniqueName: \"kubernetes.io/projected/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-kube-api-access-cnjzq\") pod \"f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc\" (UID: \"f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc\") " Aug 13 00:29:14.975658 kubelet[3169]: I0813 00:29:14.975571 3169 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-whisker-ca-bundle\") pod \"f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc\" (UID: \"f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc\") " Aug 13 00:29:14.975766 kubelet[3169]: I0813 00:29:14.975672 3169 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-whisker-backend-key-pair\") pod \"f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc\" (UID: \"f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc\") " Aug 13 00:29:14.976864 kubelet[3169]: I0813 00:29:14.976828 3169 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc" (UID: "f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Aug 13 00:29:14.977897 kubelet[3169]: I0813 00:29:14.977845 3169 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-kube-api-access-cnjzq" (OuterVolumeSpecName: "kube-api-access-cnjzq") pod "f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc" (UID: "f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc"). InnerVolumeSpecName "kube-api-access-cnjzq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Aug 13 00:29:14.980961 kubelet[3169]: I0813 00:29:14.980738 3169 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc" (UID: "f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Aug 13 00:29:15.076317 kubelet[3169]: I0813 00:29:15.076238 3169 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-cnjzq\" (UniqueName: \"kubernetes.io/projected/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-kube-api-access-cnjzq\") on node \"ci-4372.1.0-a-4f0cf7bbd9\" DevicePath \"\"" Aug 13 00:29:15.076317 kubelet[3169]: I0813 00:29:15.076264 3169 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-whisker-ca-bundle\") on node \"ci-4372.1.0-a-4f0cf7bbd9\" DevicePath \"\"" Aug 13 00:29:15.076317 kubelet[3169]: I0813 00:29:15.076274 3169 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc-whisker-backend-key-pair\") on node \"ci-4372.1.0-a-4f0cf7bbd9\" DevicePath \"\"" Aug 13 00:29:15.096158 systemd[1]: var-lib-kubelet-pods-f1e6c903\x2d9f98\x2d4e5f\x2dacbb\x2dc0e2d3c613dc-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dcnjzq.mount: Deactivated successfully. Aug 13 00:29:15.096247 systemd[1]: var-lib-kubelet-pods-f1e6c903\x2d9f98\x2d4e5f\x2dacbb\x2dc0e2d3c613dc-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Aug 13 00:29:15.147208 systemd[1]: Removed slice kubepods-besteffort-podf1e6c903_9f98_4e5f_acbb_c0e2d3c613dc.slice - libcontainer container kubepods-besteffort-podf1e6c903_9f98_4e5f_acbb_c0e2d3c613dc.slice. Aug 13 00:29:15.164763 kubelet[3169]: I0813 00:29:15.164712 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-6mkd5" podStartSLOduration=2.020916686 podStartE2EDuration="22.164700465s" podCreationTimestamp="2025-08-13 00:28:53 +0000 UTC" firstStartedPulling="2025-08-13 00:28:54.002756814 +0000 UTC m=+20.061787893" lastFinishedPulling="2025-08-13 00:29:14.146540589 +0000 UTC m=+40.205571672" observedRunningTime="2025-08-13 00:29:15.163006769 +0000 UTC m=+41.222037859" watchObservedRunningTime="2025-08-13 00:29:15.164700465 +0000 UTC m=+41.223731553" Aug 13 00:29:15.213529 containerd[1720]: time="2025-08-13T00:29:15.213492457Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3\" id:\"4092cb27967775ea0b5d3923fb24b3d1575dec87fab27b109f2b2fca90164b5e\" pid:4280 exit_status:1 exited_at:{seconds:1755044955 nanos:213274851}" Aug 13 00:29:15.238696 systemd[1]: Created slice kubepods-besteffort-pod02f5b67c_9d3e_41b9_9fb1_b3f0869e8e70.slice - libcontainer container kubepods-besteffort-pod02f5b67c_9d3e_41b9_9fb1_b3f0869e8e70.slice. Aug 13 00:29:15.276663 kubelet[3169]: I0813 00:29:15.276638 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70-whisker-backend-key-pair\") pod \"whisker-59f75767c5-s7k7q\" (UID: \"02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70\") " pod="calico-system/whisker-59f75767c5-s7k7q" Aug 13 00:29:15.276663 kubelet[3169]: I0813 00:29:15.276663 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70-whisker-ca-bundle\") pod \"whisker-59f75767c5-s7k7q\" (UID: \"02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70\") " pod="calico-system/whisker-59f75767c5-s7k7q" Aug 13 00:29:15.276758 kubelet[3169]: I0813 00:29:15.276679 3169 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6jnz\" (UniqueName: \"kubernetes.io/projected/02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70-kube-api-access-m6jnz\") pod \"whisker-59f75767c5-s7k7q\" (UID: \"02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70\") " pod="calico-system/whisker-59f75767c5-s7k7q" Aug 13 00:29:15.543880 containerd[1720]: time="2025-08-13T00:29:15.543848918Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-59f75767c5-s7k7q,Uid:02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70,Namespace:calico-system,Attempt:0,}" Aug 13 00:29:15.637382 systemd-networkd[1364]: cali6b36b2df28f: Link UP Aug 13 00:29:15.639445 systemd-networkd[1364]: cali6b36b2df28f: Gained carrier Aug 13 00:29:15.652379 containerd[1720]: 2025-08-13 00:29:15.568 [INFO][4293] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 00:29:15.652379 containerd[1720]: 2025-08-13 00:29:15.575 [INFO][4293] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0 whisker-59f75767c5- calico-system 02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70 895 0 2025-08-13 00:29:15 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:59f75767c5 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4372.1.0-a-4f0cf7bbd9 whisker-59f75767c5-s7k7q eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali6b36b2df28f [] [] }} ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Namespace="calico-system" Pod="whisker-59f75767c5-s7k7q" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-" Aug 13 00:29:15.652379 containerd[1720]: 2025-08-13 00:29:15.575 [INFO][4293] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Namespace="calico-system" Pod="whisker-59f75767c5-s7k7q" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" Aug 13 00:29:15.652379 containerd[1720]: 2025-08-13 00:29:15.593 [INFO][4304] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" HandleID="k8s-pod-network.c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" Aug 13 00:29:15.652567 containerd[1720]: 2025-08-13 00:29:15.593 [INFO][4304] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" HandleID="k8s-pod-network.c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f1c0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4372.1.0-a-4f0cf7bbd9", "pod":"whisker-59f75767c5-s7k7q", "timestamp":"2025-08-13 00:29:15.593231826 +0000 UTC"}, Hostname:"ci-4372.1.0-a-4f0cf7bbd9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:29:15.652567 containerd[1720]: 2025-08-13 00:29:15.593 [INFO][4304] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:29:15.652567 containerd[1720]: 2025-08-13 00:29:15.593 [INFO][4304] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:29:15.652567 containerd[1720]: 2025-08-13 00:29:15.593 [INFO][4304] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4372.1.0-a-4f0cf7bbd9' Aug 13 00:29:15.652567 containerd[1720]: 2025-08-13 00:29:15.597 [INFO][4304] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:15.652567 containerd[1720]: 2025-08-13 00:29:15.601 [INFO][4304] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:15.652567 containerd[1720]: 2025-08-13 00:29:15.604 [INFO][4304] ipam/ipam.go 511: Trying affinity for 192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:15.652567 containerd[1720]: 2025-08-13 00:29:15.605 [INFO][4304] ipam/ipam.go 158: Attempting to load block cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:15.652567 containerd[1720]: 2025-08-13 00:29:15.606 [INFO][4304] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:15.652749 containerd[1720]: 2025-08-13 00:29:15.606 [INFO][4304] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.61.0/26 handle="k8s-pod-network.c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:15.652749 containerd[1720]: 2025-08-13 00:29:15.607 [INFO][4304] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157 Aug 13 00:29:15.652749 containerd[1720]: 2025-08-13 00:29:15.610 [INFO][4304] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.61.0/26 handle="k8s-pod-network.c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:15.652749 containerd[1720]: 2025-08-13 00:29:15.614 [INFO][4304] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.61.1/26] block=192.168.61.0/26 handle="k8s-pod-network.c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:15.652749 containerd[1720]: 2025-08-13 00:29:15.614 [INFO][4304] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.61.1/26] handle="k8s-pod-network.c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:15.652749 containerd[1720]: 2025-08-13 00:29:15.614 [INFO][4304] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:29:15.652749 containerd[1720]: 2025-08-13 00:29:15.614 [INFO][4304] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.61.1/26] IPv6=[] ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" HandleID="k8s-pod-network.c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" Aug 13 00:29:15.652873 containerd[1720]: 2025-08-13 00:29:15.616 [INFO][4293] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Namespace="calico-system" Pod="whisker-59f75767c5-s7k7q" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0", GenerateName:"whisker-59f75767c5-", Namespace:"calico-system", SelfLink:"", UID:"02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70", ResourceVersion:"895", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 29, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"59f75767c5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"", Pod:"whisker-59f75767c5-s7k7q", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.61.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali6b36b2df28f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:15.652873 containerd[1720]: 2025-08-13 00:29:15.617 [INFO][4293] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.61.1/32] ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Namespace="calico-system" Pod="whisker-59f75767c5-s7k7q" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" Aug 13 00:29:15.652946 containerd[1720]: 2025-08-13 00:29:15.617 [INFO][4293] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6b36b2df28f ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Namespace="calico-system" Pod="whisker-59f75767c5-s7k7q" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" Aug 13 00:29:15.652946 containerd[1720]: 2025-08-13 00:29:15.637 [INFO][4293] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Namespace="calico-system" Pod="whisker-59f75767c5-s7k7q" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" Aug 13 00:29:15.652984 containerd[1720]: 2025-08-13 00:29:15.637 [INFO][4293] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Namespace="calico-system" Pod="whisker-59f75767c5-s7k7q" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0", GenerateName:"whisker-59f75767c5-", Namespace:"calico-system", SelfLink:"", UID:"02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70", ResourceVersion:"895", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 29, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"59f75767c5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157", Pod:"whisker-59f75767c5-s7k7q", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.61.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali6b36b2df28f", MAC:"12:b2:d6:cf:20:9a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:15.653029 containerd[1720]: 2025-08-13 00:29:15.650 [INFO][4293] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" Namespace="calico-system" Pod="whisker-59f75767c5-s7k7q" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-whisker--59f75767c5--s7k7q-eth0" Aug 13 00:29:15.699773 containerd[1720]: time="2025-08-13T00:29:15.699710944Z" level=info msg="connecting to shim c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157" address="unix:///run/containerd/s/61a8b048e40ee6205a6bb1a1c4187191c12b2213cd5cc6ff53a466b245577e93" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:29:15.718421 systemd[1]: Started cri-containerd-c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157.scope - libcontainer container c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157. Aug 13 00:29:15.756116 containerd[1720]: time="2025-08-13T00:29:15.756089143Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-59f75767c5-s7k7q,Uid:02f5b67c-9d3e-41b9-9fb1-b3f0869e8e70,Namespace:calico-system,Attempt:0,} returns sandbox id \"c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157\"" Aug 13 00:29:15.757366 containerd[1720]: time="2025-08-13T00:29:15.757344192Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\"" Aug 13 00:29:16.023922 kubelet[3169]: I0813 00:29:16.023891 3169 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc" path="/var/lib/kubelet/pods/f1e6c903-9f98-4e5f-acbb-c0e2d3c613dc/volumes" Aug 13 00:29:16.352881 containerd[1720]: time="2025-08-13T00:29:16.352795008Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3\" id:\"48ddd6153b80e67e53e2240d754c033513917a48f4f343d5c551da8c8ac2218c\" pid:4463 exit_status:1 exited_at:{seconds:1755044956 nanos:352046734}" Aug 13 00:29:16.662903 systemd-networkd[1364]: vxlan.calico: Link UP Aug 13 00:29:16.662918 systemd-networkd[1364]: vxlan.calico: Gained carrier Aug 13 00:29:17.022490 containerd[1720]: time="2025-08-13T00:29:17.022459187Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-4c59z,Uid:eaaa778d-497a-49e1-b653-c6dedc50adee,Namespace:calico-system,Attempt:0,}" Aug 13 00:29:17.115929 systemd-networkd[1364]: cali7d55374243c: Link UP Aug 13 00:29:17.116239 systemd-networkd[1364]: cali7d55374243c: Gained carrier Aug 13 00:29:17.129933 containerd[1720]: 2025-08-13 00:29:17.058 [INFO][4578] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0 goldmane-768f4c5c69- calico-system eaaa778d-497a-49e1-b653-c6dedc50adee 826 0 2025-08-13 00:28:52 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:768f4c5c69 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4372.1.0-a-4f0cf7bbd9 goldmane-768f4c5c69-4c59z eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali7d55374243c [] [] }} ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Namespace="calico-system" Pod="goldmane-768f4c5c69-4c59z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-" Aug 13 00:29:17.129933 containerd[1720]: 2025-08-13 00:29:17.058 [INFO][4578] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Namespace="calico-system" Pod="goldmane-768f4c5c69-4c59z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" Aug 13 00:29:17.129933 containerd[1720]: 2025-08-13 00:29:17.080 [INFO][4590] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" HandleID="k8s-pod-network.c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" Aug 13 00:29:17.130166 containerd[1720]: 2025-08-13 00:29:17.080 [INFO][4590] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" HandleID="k8s-pod-network.c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5680), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4372.1.0-a-4f0cf7bbd9", "pod":"goldmane-768f4c5c69-4c59z", "timestamp":"2025-08-13 00:29:17.080162158 +0000 UTC"}, Hostname:"ci-4372.1.0-a-4f0cf7bbd9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:29:17.130166 containerd[1720]: 2025-08-13 00:29:17.080 [INFO][4590] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:29:17.130166 containerd[1720]: 2025-08-13 00:29:17.080 [INFO][4590] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:29:17.130166 containerd[1720]: 2025-08-13 00:29:17.080 [INFO][4590] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4372.1.0-a-4f0cf7bbd9' Aug 13 00:29:17.130166 containerd[1720]: 2025-08-13 00:29:17.086 [INFO][4590] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:17.130166 containerd[1720]: 2025-08-13 00:29:17.090 [INFO][4590] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:17.130166 containerd[1720]: 2025-08-13 00:29:17.093 [INFO][4590] ipam/ipam.go 511: Trying affinity for 192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:17.130166 containerd[1720]: 2025-08-13 00:29:17.095 [INFO][4590] ipam/ipam.go 158: Attempting to load block cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:17.130166 containerd[1720]: 2025-08-13 00:29:17.097 [INFO][4590] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:17.130993 containerd[1720]: 2025-08-13 00:29:17.097 [INFO][4590] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.61.0/26 handle="k8s-pod-network.c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:17.130993 containerd[1720]: 2025-08-13 00:29:17.099 [INFO][4590] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee Aug 13 00:29:17.130993 containerd[1720]: 2025-08-13 00:29:17.103 [INFO][4590] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.61.0/26 handle="k8s-pod-network.c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:17.130993 containerd[1720]: 2025-08-13 00:29:17.112 [INFO][4590] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.61.2/26] block=192.168.61.0/26 handle="k8s-pod-network.c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:17.130993 containerd[1720]: 2025-08-13 00:29:17.112 [INFO][4590] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.61.2/26] handle="k8s-pod-network.c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:17.130993 containerd[1720]: 2025-08-13 00:29:17.112 [INFO][4590] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:29:17.130993 containerd[1720]: 2025-08-13 00:29:17.112 [INFO][4590] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.61.2/26] IPv6=[] ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" HandleID="k8s-pod-network.c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" Aug 13 00:29:17.131148 containerd[1720]: 2025-08-13 00:29:17.113 [INFO][4578] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Namespace="calico-system" Pod="goldmane-768f4c5c69-4c59z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"eaaa778d-497a-49e1-b653-c6dedc50adee", ResourceVersion:"826", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"", Pod:"goldmane-768f4c5c69-4c59z", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.61.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7d55374243c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:17.131148 containerd[1720]: 2025-08-13 00:29:17.113 [INFO][4578] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.61.2/32] ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Namespace="calico-system" Pod="goldmane-768f4c5c69-4c59z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" Aug 13 00:29:17.131233 containerd[1720]: 2025-08-13 00:29:17.113 [INFO][4578] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7d55374243c ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Namespace="calico-system" Pod="goldmane-768f4c5c69-4c59z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" Aug 13 00:29:17.131233 containerd[1720]: 2025-08-13 00:29:17.115 [INFO][4578] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Namespace="calico-system" Pod="goldmane-768f4c5c69-4c59z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" Aug 13 00:29:17.131280 containerd[1720]: 2025-08-13 00:29:17.115 [INFO][4578] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Namespace="calico-system" Pod="goldmane-768f4c5c69-4c59z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"eaaa778d-497a-49e1-b653-c6dedc50adee", ResourceVersion:"826", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee", Pod:"goldmane-768f4c5c69-4c59z", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.61.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali7d55374243c", MAC:"9a:54:30:dc:6e:07", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:17.131358 containerd[1720]: 2025-08-13 00:29:17.125 [INFO][4578] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" Namespace="calico-system" Pod="goldmane-768f4c5c69-4c59z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-goldmane--768f4c5c69--4c59z-eth0" Aug 13 00:29:17.183503 containerd[1720]: time="2025-08-13T00:29:17.183444381Z" level=info msg="connecting to shim c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee" address="unix:///run/containerd/s/79ddc319d98a8ed791990edb52f1f39cfc75790eb2c4ead8d4c7ff27ee7c823d" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:29:17.209609 systemd[1]: Started cri-containerd-c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee.scope - libcontainer container c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee. Aug 13 00:29:17.252869 containerd[1720]: time="2025-08-13T00:29:17.252845876Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-4c59z,Uid:eaaa778d-497a-49e1-b653-c6dedc50adee,Namespace:calico-system,Attempt:0,} returns sandbox id \"c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee\"" Aug 13 00:29:17.254873 containerd[1720]: time="2025-08-13T00:29:17.254848891Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:17.257412 containerd[1720]: time="2025-08-13T00:29:17.257386889Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.2: active requests=0, bytes read=4661207" Aug 13 00:29:17.260740 containerd[1720]: time="2025-08-13T00:29:17.260696707Z" level=info msg="ImageCreate event name:\"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:17.264134 containerd[1720]: time="2025-08-13T00:29:17.264096885Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:17.264650 containerd[1720]: time="2025-08-13T00:29:17.264416006Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.2\" with image id \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\", size \"6153902\" in 1.507041175s" Aug 13 00:29:17.264650 containerd[1720]: time="2025-08-13T00:29:17.264440680Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\" returns image reference \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\"" Aug 13 00:29:17.266100 containerd[1720]: time="2025-08-13T00:29:17.265190136Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\"" Aug 13 00:29:17.270232 containerd[1720]: time="2025-08-13T00:29:17.270203422Z" level=info msg="CreateContainer within sandbox \"c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Aug 13 00:29:17.284047 containerd[1720]: time="2025-08-13T00:29:17.283992125Z" level=info msg="Container 9249c9c0f5ae21d1fc06834cfb6a74cec4487f92adc2d8493c7af84158d14fa8: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:17.297147 containerd[1720]: time="2025-08-13T00:29:17.297125196Z" level=info msg="CreateContainer within sandbox \"c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"9249c9c0f5ae21d1fc06834cfb6a74cec4487f92adc2d8493c7af84158d14fa8\"" Aug 13 00:29:17.297563 containerd[1720]: time="2025-08-13T00:29:17.297539210Z" level=info msg="StartContainer for \"9249c9c0f5ae21d1fc06834cfb6a74cec4487f92adc2d8493c7af84158d14fa8\"" Aug 13 00:29:17.298612 containerd[1720]: time="2025-08-13T00:29:17.298589472Z" level=info msg="connecting to shim 9249c9c0f5ae21d1fc06834cfb6a74cec4487f92adc2d8493c7af84158d14fa8" address="unix:///run/containerd/s/61a8b048e40ee6205a6bb1a1c4187191c12b2213cd5cc6ff53a466b245577e93" protocol=ttrpc version=3 Aug 13 00:29:17.313416 systemd[1]: Started cri-containerd-9249c9c0f5ae21d1fc06834cfb6a74cec4487f92adc2d8493c7af84158d14fa8.scope - libcontainer container 9249c9c0f5ae21d1fc06834cfb6a74cec4487f92adc2d8493c7af84158d14fa8. Aug 13 00:29:17.348118 containerd[1720]: time="2025-08-13T00:29:17.348077404Z" level=info msg="StartContainer for \"9249c9c0f5ae21d1fc06834cfb6a74cec4487f92adc2d8493c7af84158d14fa8\" returns successfully" Aug 13 00:29:17.530412 systemd-networkd[1364]: cali6b36b2df28f: Gained IPv6LL Aug 13 00:29:18.024434 containerd[1720]: time="2025-08-13T00:29:18.024387900Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d685458f5-fchq5,Uid:3243471f-918a-41b7-bcb5-e0d349b40c90,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:29:18.042505 systemd-networkd[1364]: vxlan.calico: Gained IPv6LL Aug 13 00:29:18.116034 systemd-networkd[1364]: caliac62f891ba8: Link UP Aug 13 00:29:18.116924 systemd-networkd[1364]: caliac62f891ba8: Gained carrier Aug 13 00:29:18.128985 containerd[1720]: 2025-08-13 00:29:18.064 [INFO][4690] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0 calico-apiserver-7d685458f5- calico-apiserver 3243471f-918a-41b7-bcb5-e0d349b40c90 825 0 2025-08-13 00:28:49 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7d685458f5 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4372.1.0-a-4f0cf7bbd9 calico-apiserver-7d685458f5-fchq5 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] caliac62f891ba8 [] [] }} ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-fchq5" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-" Aug 13 00:29:18.128985 containerd[1720]: 2025-08-13 00:29:18.064 [INFO][4690] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-fchq5" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" Aug 13 00:29:18.128985 containerd[1720]: 2025-08-13 00:29:18.083 [INFO][4702] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" HandleID="k8s-pod-network.14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" Aug 13 00:29:18.129171 containerd[1720]: 2025-08-13 00:29:18.084 [INFO][4702] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" HandleID="k8s-pod-network.14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5040), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4372.1.0-a-4f0cf7bbd9", "pod":"calico-apiserver-7d685458f5-fchq5", "timestamp":"2025-08-13 00:29:18.083874903 +0000 UTC"}, Hostname:"ci-4372.1.0-a-4f0cf7bbd9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:29:18.129171 containerd[1720]: 2025-08-13 00:29:18.084 [INFO][4702] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:29:18.129171 containerd[1720]: 2025-08-13 00:29:18.084 [INFO][4702] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:29:18.129171 containerd[1720]: 2025-08-13 00:29:18.084 [INFO][4702] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4372.1.0-a-4f0cf7bbd9' Aug 13 00:29:18.129171 containerd[1720]: 2025-08-13 00:29:18.088 [INFO][4702] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:18.129171 containerd[1720]: 2025-08-13 00:29:18.091 [INFO][4702] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:18.129171 containerd[1720]: 2025-08-13 00:29:18.095 [INFO][4702] ipam/ipam.go 511: Trying affinity for 192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:18.129171 containerd[1720]: 2025-08-13 00:29:18.097 [INFO][4702] ipam/ipam.go 158: Attempting to load block cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:18.129171 containerd[1720]: 2025-08-13 00:29:18.098 [INFO][4702] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:18.129705 containerd[1720]: 2025-08-13 00:29:18.098 [INFO][4702] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.61.0/26 handle="k8s-pod-network.14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:18.129705 containerd[1720]: 2025-08-13 00:29:18.099 [INFO][4702] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544 Aug 13 00:29:18.129705 containerd[1720]: 2025-08-13 00:29:18.103 [INFO][4702] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.61.0/26 handle="k8s-pod-network.14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:18.129705 containerd[1720]: 2025-08-13 00:29:18.112 [INFO][4702] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.61.3/26] block=192.168.61.0/26 handle="k8s-pod-network.14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:18.129705 containerd[1720]: 2025-08-13 00:29:18.112 [INFO][4702] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.61.3/26] handle="k8s-pod-network.14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:18.129705 containerd[1720]: 2025-08-13 00:29:18.112 [INFO][4702] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:29:18.129705 containerd[1720]: 2025-08-13 00:29:18.112 [INFO][4702] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.61.3/26] IPv6=[] ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" HandleID="k8s-pod-network.14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" Aug 13 00:29:18.129959 containerd[1720]: 2025-08-13 00:29:18.113 [INFO][4690] cni-plugin/k8s.go 418: Populated endpoint ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-fchq5" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0", GenerateName:"calico-apiserver-7d685458f5-", Namespace:"calico-apiserver", SelfLink:"", UID:"3243471f-918a-41b7-bcb5-e0d349b40c90", ResourceVersion:"825", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7d685458f5", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"", Pod:"calico-apiserver-7d685458f5-fchq5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.61.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliac62f891ba8", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:18.130056 containerd[1720]: 2025-08-13 00:29:18.113 [INFO][4690] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.61.3/32] ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-fchq5" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" Aug 13 00:29:18.130056 containerd[1720]: 2025-08-13 00:29:18.113 [INFO][4690] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliac62f891ba8 ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-fchq5" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" Aug 13 00:29:18.130056 containerd[1720]: 2025-08-13 00:29:18.117 [INFO][4690] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-fchq5" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" Aug 13 00:29:18.130120 containerd[1720]: 2025-08-13 00:29:18.117 [INFO][4690] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-fchq5" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0", GenerateName:"calico-apiserver-7d685458f5-", Namespace:"calico-apiserver", SelfLink:"", UID:"3243471f-918a-41b7-bcb5-e0d349b40c90", ResourceVersion:"825", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7d685458f5", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544", Pod:"calico-apiserver-7d685458f5-fchq5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.61.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliac62f891ba8", MAC:"fe:ee:6f:51:9c:a9", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:18.130174 containerd[1720]: 2025-08-13 00:29:18.125 [INFO][4690] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-fchq5" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--fchq5-eth0" Aug 13 00:29:18.180024 containerd[1720]: time="2025-08-13T00:29:18.179605906Z" level=info msg="connecting to shim 14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544" address="unix:///run/containerd/s/9d0e8d62cec494e81271534a2ad51cada0189d0e9bf9f61ea6cfd941664de1b1" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:29:18.200427 systemd[1]: Started cri-containerd-14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544.scope - libcontainer container 14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544. Aug 13 00:29:18.238784 containerd[1720]: time="2025-08-13T00:29:18.238761682Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d685458f5-fchq5,Uid:3243471f-918a-41b7-bcb5-e0d349b40c90,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544\"" Aug 13 00:29:18.810402 systemd-networkd[1364]: cali7d55374243c: Gained IPv6LL Aug 13 00:29:19.022578 containerd[1720]: time="2025-08-13T00:29:19.022480311Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-nbpw8,Uid:71824225-ae9f-4348-bcd0-e9d39f3bdee6,Namespace:kube-system,Attempt:0,}" Aug 13 00:29:19.022967 containerd[1720]: time="2025-08-13T00:29:19.022505342Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-ptpbv,Uid:a62490c7-241c-47fd-bc6c-9b6d06fb91b7,Namespace:calico-system,Attempt:0,}" Aug 13 00:29:19.137014 systemd-networkd[1364]: cali93a4306afc6: Link UP Aug 13 00:29:19.137184 systemd-networkd[1364]: cali93a4306afc6: Gained carrier Aug 13 00:29:19.154988 containerd[1720]: 2025-08-13 00:29:19.066 [INFO][4767] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0 coredns-674b8bbfcf- kube-system 71824225-ae9f-4348-bcd0-e9d39f3bdee6 823 0 2025-08-13 00:28:40 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4372.1.0-a-4f0cf7bbd9 coredns-674b8bbfcf-nbpw8 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali93a4306afc6 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Namespace="kube-system" Pod="coredns-674b8bbfcf-nbpw8" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-" Aug 13 00:29:19.154988 containerd[1720]: 2025-08-13 00:29:19.066 [INFO][4767] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Namespace="kube-system" Pod="coredns-674b8bbfcf-nbpw8" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" Aug 13 00:29:19.154988 containerd[1720]: 2025-08-13 00:29:19.101 [INFO][4792] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" HandleID="k8s-pod-network.fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" Aug 13 00:29:19.155524 containerd[1720]: 2025-08-13 00:29:19.101 [INFO][4792] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" HandleID="k8s-pod-network.fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004e100), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4372.1.0-a-4f0cf7bbd9", "pod":"coredns-674b8bbfcf-nbpw8", "timestamp":"2025-08-13 00:29:19.101022717 +0000 UTC"}, Hostname:"ci-4372.1.0-a-4f0cf7bbd9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:29:19.155524 containerd[1720]: 2025-08-13 00:29:19.101 [INFO][4792] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:29:19.155524 containerd[1720]: 2025-08-13 00:29:19.101 [INFO][4792] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:29:19.155524 containerd[1720]: 2025-08-13 00:29:19.101 [INFO][4792] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4372.1.0-a-4f0cf7bbd9' Aug 13 00:29:19.155524 containerd[1720]: 2025-08-13 00:29:19.107 [INFO][4792] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.155524 containerd[1720]: 2025-08-13 00:29:19.111 [INFO][4792] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.155524 containerd[1720]: 2025-08-13 00:29:19.117 [INFO][4792] ipam/ipam.go 511: Trying affinity for 192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.155524 containerd[1720]: 2025-08-13 00:29:19.119 [INFO][4792] ipam/ipam.go 158: Attempting to load block cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.155524 containerd[1720]: 2025-08-13 00:29:19.121 [INFO][4792] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.155749 containerd[1720]: 2025-08-13 00:29:19.121 [INFO][4792] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.61.0/26 handle="k8s-pod-network.fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.155749 containerd[1720]: 2025-08-13 00:29:19.122 [INFO][4792] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb Aug 13 00:29:19.155749 containerd[1720]: 2025-08-13 00:29:19.125 [INFO][4792] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.61.0/26 handle="k8s-pod-network.fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.155749 containerd[1720]: 2025-08-13 00:29:19.131 [INFO][4792] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.61.4/26] block=192.168.61.0/26 handle="k8s-pod-network.fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.155749 containerd[1720]: 2025-08-13 00:29:19.132 [INFO][4792] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.61.4/26] handle="k8s-pod-network.fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.155749 containerd[1720]: 2025-08-13 00:29:19.132 [INFO][4792] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:29:19.155749 containerd[1720]: 2025-08-13 00:29:19.132 [INFO][4792] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.61.4/26] IPv6=[] ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" HandleID="k8s-pod-network.fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" Aug 13 00:29:19.155905 containerd[1720]: 2025-08-13 00:29:19.133 [INFO][4767] cni-plugin/k8s.go 418: Populated endpoint ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Namespace="kube-system" Pod="coredns-674b8bbfcf-nbpw8" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"71824225-ae9f-4348-bcd0-e9d39f3bdee6", ResourceVersion:"823", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"", Pod:"coredns-674b8bbfcf-nbpw8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.61.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali93a4306afc6", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:19.155905 containerd[1720]: 2025-08-13 00:29:19.133 [INFO][4767] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.61.4/32] ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Namespace="kube-system" Pod="coredns-674b8bbfcf-nbpw8" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" Aug 13 00:29:19.155905 containerd[1720]: 2025-08-13 00:29:19.133 [INFO][4767] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali93a4306afc6 ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Namespace="kube-system" Pod="coredns-674b8bbfcf-nbpw8" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" Aug 13 00:29:19.155905 containerd[1720]: 2025-08-13 00:29:19.138 [INFO][4767] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Namespace="kube-system" Pod="coredns-674b8bbfcf-nbpw8" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" Aug 13 00:29:19.155905 containerd[1720]: 2025-08-13 00:29:19.139 [INFO][4767] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Namespace="kube-system" Pod="coredns-674b8bbfcf-nbpw8" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"71824225-ae9f-4348-bcd0-e9d39f3bdee6", ResourceVersion:"823", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb", Pod:"coredns-674b8bbfcf-nbpw8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.61.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali93a4306afc6", MAC:"0e:43:f1:e9:e2:f2", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:19.155905 containerd[1720]: 2025-08-13 00:29:19.153 [INFO][4767] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" Namespace="kube-system" Pod="coredns-674b8bbfcf-nbpw8" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--nbpw8-eth0" Aug 13 00:29:19.240951 systemd-networkd[1364]: calic2edabfc539: Link UP Aug 13 00:29:19.241742 systemd-networkd[1364]: calic2edabfc539: Gained carrier Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.081 [INFO][4778] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0 csi-node-driver- calico-system a62490c7-241c-47fd-bc6c-9b6d06fb91b7 705 0 2025-08-13 00:28:53 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:8967bcb6f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4372.1.0-a-4f0cf7bbd9 csi-node-driver-ptpbv eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calic2edabfc539 [] [] }} ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Namespace="calico-system" Pod="csi-node-driver-ptpbv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.081 [INFO][4778] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Namespace="calico-system" Pod="csi-node-driver-ptpbv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.119 [INFO][4797] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" HandleID="k8s-pod-network.5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.120 [INFO][4797] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" HandleID="k8s-pod-network.5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d56a0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4372.1.0-a-4f0cf7bbd9", "pod":"csi-node-driver-ptpbv", "timestamp":"2025-08-13 00:29:19.119920467 +0000 UTC"}, Hostname:"ci-4372.1.0-a-4f0cf7bbd9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.120 [INFO][4797] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.132 [INFO][4797] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.132 [INFO][4797] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4372.1.0-a-4f0cf7bbd9' Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.207 [INFO][4797] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.211 [INFO][4797] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.217 [INFO][4797] ipam/ipam.go 511: Trying affinity for 192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.218 [INFO][4797] ipam/ipam.go 158: Attempting to load block cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.220 [INFO][4797] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.220 [INFO][4797] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.61.0/26 handle="k8s-pod-network.5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.222 [INFO][4797] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91 Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.228 [INFO][4797] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.61.0/26 handle="k8s-pod-network.5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.234 [INFO][4797] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.61.5/26] block=192.168.61.0/26 handle="k8s-pod-network.5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.234 [INFO][4797] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.61.5/26] handle="k8s-pod-network.5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.234 [INFO][4797] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:29:19.254092 containerd[1720]: 2025-08-13 00:29:19.234 [INFO][4797] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.61.5/26] IPv6=[] ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" HandleID="k8s-pod-network.5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" Aug 13 00:29:19.255392 containerd[1720]: 2025-08-13 00:29:19.236 [INFO][4778] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Namespace="calico-system" Pod="csi-node-driver-ptpbv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a62490c7-241c-47fd-bc6c-9b6d06fb91b7", ResourceVersion:"705", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"", Pod:"csi-node-driver-ptpbv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.61.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic2edabfc539", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:19.255392 containerd[1720]: 2025-08-13 00:29:19.236 [INFO][4778] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.61.5/32] ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Namespace="calico-system" Pod="csi-node-driver-ptpbv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" Aug 13 00:29:19.255392 containerd[1720]: 2025-08-13 00:29:19.236 [INFO][4778] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic2edabfc539 ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Namespace="calico-system" Pod="csi-node-driver-ptpbv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" Aug 13 00:29:19.255392 containerd[1720]: 2025-08-13 00:29:19.240 [INFO][4778] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Namespace="calico-system" Pod="csi-node-driver-ptpbv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" Aug 13 00:29:19.255392 containerd[1720]: 2025-08-13 00:29:19.241 [INFO][4778] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Namespace="calico-system" Pod="csi-node-driver-ptpbv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a62490c7-241c-47fd-bc6c-9b6d06fb91b7", ResourceVersion:"705", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91", Pod:"csi-node-driver-ptpbv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.61.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic2edabfc539", MAC:"0e:2a:99:ea:47:ba", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:19.255392 containerd[1720]: 2025-08-13 00:29:19.251 [INFO][4778] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" Namespace="calico-system" Pod="csi-node-driver-ptpbv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-csi--node--driver--ptpbv-eth0" Aug 13 00:29:19.322377 systemd-networkd[1364]: caliac62f891ba8: Gained IPv6LL Aug 13 00:29:19.521240 containerd[1720]: time="2025-08-13T00:29:19.521172274Z" level=info msg="connecting to shim 5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91" address="unix:///run/containerd/s/5bb2fcb84cce5ed65c00c6b11ab1459b160b63a7349b49e235338634855ab9d9" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:29:19.524602 containerd[1720]: time="2025-08-13T00:29:19.524565723Z" level=info msg="connecting to shim fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb" address="unix:///run/containerd/s/6ef19e956c48e85cae844c5a19cffb984a666d608ea54c0d97cc92346171b0b6" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:29:19.563506 systemd[1]: Started cri-containerd-5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91.scope - libcontainer container 5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91. Aug 13 00:29:19.568623 systemd[1]: Started cri-containerd-fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb.scope - libcontainer container fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb. Aug 13 00:29:19.624805 containerd[1720]: time="2025-08-13T00:29:19.624778879Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-ptpbv,Uid:a62490c7-241c-47fd-bc6c-9b6d06fb91b7,Namespace:calico-system,Attempt:0,} returns sandbox id \"5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91\"" Aug 13 00:29:19.644559 containerd[1720]: time="2025-08-13T00:29:19.644535097Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-nbpw8,Uid:71824225-ae9f-4348-bcd0-e9d39f3bdee6,Namespace:kube-system,Attempt:0,} returns sandbox id \"fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb\"" Aug 13 00:29:19.651724 containerd[1720]: time="2025-08-13T00:29:19.651704484Z" level=info msg="CreateContainer within sandbox \"fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 00:29:19.672608 containerd[1720]: time="2025-08-13T00:29:19.672585838Z" level=info msg="Container 344623241bc5257354e24f6350a85e32e8d8967322f201baac5908c899b91987: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:19.683618 containerd[1720]: time="2025-08-13T00:29:19.683597960Z" level=info msg="CreateContainer within sandbox \"fc5a2b4ee3ba1a6fcf7185fa5220cd60df940f7c6a1c7eaa9917baab8405d7cb\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"344623241bc5257354e24f6350a85e32e8d8967322f201baac5908c899b91987\"" Aug 13 00:29:19.685569 containerd[1720]: time="2025-08-13T00:29:19.685539952Z" level=info msg="StartContainer for \"344623241bc5257354e24f6350a85e32e8d8967322f201baac5908c899b91987\"" Aug 13 00:29:19.686721 containerd[1720]: time="2025-08-13T00:29:19.686694671Z" level=info msg="connecting to shim 344623241bc5257354e24f6350a85e32e8d8967322f201baac5908c899b91987" address="unix:///run/containerd/s/6ef19e956c48e85cae844c5a19cffb984a666d608ea54c0d97cc92346171b0b6" protocol=ttrpc version=3 Aug 13 00:29:19.703508 systemd[1]: Started cri-containerd-344623241bc5257354e24f6350a85e32e8d8967322f201baac5908c899b91987.scope - libcontainer container 344623241bc5257354e24f6350a85e32e8d8967322f201baac5908c899b91987. Aug 13 00:29:19.739572 containerd[1720]: time="2025-08-13T00:29:19.739548052Z" level=info msg="StartContainer for \"344623241bc5257354e24f6350a85e32e8d8967322f201baac5908c899b91987\" returns successfully" Aug 13 00:29:20.024254 containerd[1720]: time="2025-08-13T00:29:20.023992293Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d685458f5-h9nsv,Uid:f36fa87a-1607-484a-99a0-3c120173b24c,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:29:20.178344 kubelet[3169]: I0813 00:29:20.178223 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-nbpw8" podStartSLOduration=40.177857417 podStartE2EDuration="40.177857417s" podCreationTimestamp="2025-08-13 00:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:29:20.177572415 +0000 UTC m=+46.236603500" watchObservedRunningTime="2025-08-13 00:29:20.177857417 +0000 UTC m=+46.236888510" Aug 13 00:29:20.196646 systemd-networkd[1364]: calidf6d7645133: Link UP Aug 13 00:29:20.198847 systemd-networkd[1364]: calidf6d7645133: Gained carrier Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.093 [INFO][4958] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0 calico-apiserver-7d685458f5- calico-apiserver f36fa87a-1607-484a-99a0-3c120173b24c 828 0 2025-08-13 00:28:49 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7d685458f5 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4372.1.0-a-4f0cf7bbd9 calico-apiserver-7d685458f5-h9nsv eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calidf6d7645133 [] [] }} ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-h9nsv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.093 [INFO][4958] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-h9nsv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.142 [INFO][4973] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" HandleID="k8s-pod-network.8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.142 [INFO][4973] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" HandleID="k8s-pod-network.8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0006100d0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4372.1.0-a-4f0cf7bbd9", "pod":"calico-apiserver-7d685458f5-h9nsv", "timestamp":"2025-08-13 00:29:20.142234167 +0000 UTC"}, Hostname:"ci-4372.1.0-a-4f0cf7bbd9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.142 [INFO][4973] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.142 [INFO][4973] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.142 [INFO][4973] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4372.1.0-a-4f0cf7bbd9' Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.149 [INFO][4973] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.152 [INFO][4973] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.156 [INFO][4973] ipam/ipam.go 511: Trying affinity for 192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.160 [INFO][4973] ipam/ipam.go 158: Attempting to load block cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.163 [INFO][4973] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.163 [INFO][4973] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.61.0/26 handle="k8s-pod-network.8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.165 [INFO][4973] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.173 [INFO][4973] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.61.0/26 handle="k8s-pod-network.8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.184 [INFO][4973] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.61.6/26] block=192.168.61.0/26 handle="k8s-pod-network.8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.184 [INFO][4973] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.61.6/26] handle="k8s-pod-network.8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.184 [INFO][4973] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:29:20.233310 containerd[1720]: 2025-08-13 00:29:20.184 [INFO][4973] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.61.6/26] IPv6=[] ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" HandleID="k8s-pod-network.8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" Aug 13 00:29:20.234636 containerd[1720]: 2025-08-13 00:29:20.190 [INFO][4958] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-h9nsv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0", GenerateName:"calico-apiserver-7d685458f5-", Namespace:"calico-apiserver", SelfLink:"", UID:"f36fa87a-1607-484a-99a0-3c120173b24c", ResourceVersion:"828", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7d685458f5", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"", Pod:"calico-apiserver-7d685458f5-h9nsv", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.61.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidf6d7645133", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:20.234636 containerd[1720]: 2025-08-13 00:29:20.190 [INFO][4958] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.61.6/32] ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-h9nsv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" Aug 13 00:29:20.234636 containerd[1720]: 2025-08-13 00:29:20.190 [INFO][4958] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calidf6d7645133 ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-h9nsv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" Aug 13 00:29:20.234636 containerd[1720]: 2025-08-13 00:29:20.203 [INFO][4958] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-h9nsv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" Aug 13 00:29:20.234636 containerd[1720]: 2025-08-13 00:29:20.207 [INFO][4958] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-h9nsv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0", GenerateName:"calico-apiserver-7d685458f5-", Namespace:"calico-apiserver", SelfLink:"", UID:"f36fa87a-1607-484a-99a0-3c120173b24c", ResourceVersion:"828", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7d685458f5", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f", Pod:"calico-apiserver-7d685458f5-h9nsv", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.61.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidf6d7645133", MAC:"16:55:cd:2d:12:f4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:20.234636 containerd[1720]: 2025-08-13 00:29:20.230 [INFO][4958] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" Namespace="calico-apiserver" Pod="calico-apiserver-7d685458f5-h9nsv" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--apiserver--7d685458f5--h9nsv-eth0" Aug 13 00:29:20.285354 containerd[1720]: time="2025-08-13T00:29:20.285105338Z" level=info msg="connecting to shim 8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f" address="unix:///run/containerd/s/8b60eb06c6910877f06d48a17bdc5feba86c9cc888921f90d48172146eba1928" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:29:20.312487 systemd[1]: Started cri-containerd-8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f.scope - libcontainer container 8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f. Aug 13 00:29:20.373791 containerd[1720]: time="2025-08-13T00:29:20.373761154Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d685458f5-h9nsv,Uid:f36fa87a-1607-484a-99a0-3c120173b24c,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f\"" Aug 13 00:29:20.410635 systemd-networkd[1364]: calic2edabfc539: Gained IPv6LL Aug 13 00:29:20.487637 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1486265531.mount: Deactivated successfully. Aug 13 00:29:20.573248 containerd[1720]: time="2025-08-13T00:29:20.573182781Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:20.575366 containerd[1720]: time="2025-08-13T00:29:20.575341265Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.2: active requests=0, bytes read=66352308" Aug 13 00:29:20.580258 containerd[1720]: time="2025-08-13T00:29:20.580218754Z" level=info msg="ImageCreate event name:\"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:20.585450 containerd[1720]: time="2025-08-13T00:29:20.585410441Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:20.586071 containerd[1720]: time="2025-08-13T00:29:20.585821123Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" with image id \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\", size \"66352154\" in 3.320604001s" Aug 13 00:29:20.586071 containerd[1720]: time="2025-08-13T00:29:20.585847687Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" returns image reference \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\"" Aug 13 00:29:20.586604 containerd[1720]: time="2025-08-13T00:29:20.586589905Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\"" Aug 13 00:29:20.592703 containerd[1720]: time="2025-08-13T00:29:20.592679445Z" level=info msg="CreateContainer within sandbox \"c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Aug 13 00:29:20.612306 containerd[1720]: time="2025-08-13T00:29:20.610928244Z" level=info msg="Container ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:20.625305 containerd[1720]: time="2025-08-13T00:29:20.625262579Z" level=info msg="CreateContainer within sandbox \"c9d4f9f46aa4ee7794052180a137fc65078c001c7e6fbf21c613c0bee740bcee\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\"" Aug 13 00:29:20.626473 containerd[1720]: time="2025-08-13T00:29:20.625818012Z" level=info msg="StartContainer for \"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\"" Aug 13 00:29:20.626762 containerd[1720]: time="2025-08-13T00:29:20.626740225Z" level=info msg="connecting to shim ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943" address="unix:///run/containerd/s/79ddc319d98a8ed791990edb52f1f39cfc75790eb2c4ead8d4c7ff27ee7c823d" protocol=ttrpc version=3 Aug 13 00:29:20.650518 systemd[1]: Started cri-containerd-ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943.scope - libcontainer container ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943. Aug 13 00:29:20.707370 containerd[1720]: time="2025-08-13T00:29:20.707306106Z" level=info msg="StartContainer for \"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\" returns successfully" Aug 13 00:29:20.794530 systemd-networkd[1364]: cali93a4306afc6: Gained IPv6LL Aug 13 00:29:21.022682 containerd[1720]: time="2025-08-13T00:29:21.022642060Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hhdbb,Uid:2680547d-4498-4eac-9daa-64c89f065d86,Namespace:kube-system,Attempt:0,}" Aug 13 00:29:21.107420 systemd-networkd[1364]: cali13161694bfa: Link UP Aug 13 00:29:21.108123 systemd-networkd[1364]: cali13161694bfa: Gained carrier Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.055 [INFO][5072] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0 coredns-674b8bbfcf- kube-system 2680547d-4498-4eac-9daa-64c89f065d86 822 0 2025-08-13 00:28:40 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4372.1.0-a-4f0cf7bbd9 coredns-674b8bbfcf-hhdbb eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali13161694bfa [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Namespace="kube-system" Pod="coredns-674b8bbfcf-hhdbb" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.055 [INFO][5072] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Namespace="kube-system" Pod="coredns-674b8bbfcf-hhdbb" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.073 [INFO][5084] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" HandleID="k8s-pod-network.57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.074 [INFO][5084] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" HandleID="k8s-pod-network.57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5960), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4372.1.0-a-4f0cf7bbd9", "pod":"coredns-674b8bbfcf-hhdbb", "timestamp":"2025-08-13 00:29:21.073954014 +0000 UTC"}, Hostname:"ci-4372.1.0-a-4f0cf7bbd9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.074 [INFO][5084] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.074 [INFO][5084] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.074 [INFO][5084] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4372.1.0-a-4f0cf7bbd9' Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.077 [INFO][5084] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.081 [INFO][5084] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.087 [INFO][5084] ipam/ipam.go 511: Trying affinity for 192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.088 [INFO][5084] ipam/ipam.go 158: Attempting to load block cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.090 [INFO][5084] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.090 [INFO][5084] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.61.0/26 handle="k8s-pod-network.57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.091 [INFO][5084] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.095 [INFO][5084] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.61.0/26 handle="k8s-pod-network.57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.102 [INFO][5084] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.61.7/26] block=192.168.61.0/26 handle="k8s-pod-network.57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.102 [INFO][5084] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.61.7/26] handle="k8s-pod-network.57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.102 [INFO][5084] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:29:21.121304 containerd[1720]: 2025-08-13 00:29:21.102 [INFO][5084] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.61.7/26] IPv6=[] ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" HandleID="k8s-pod-network.57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" Aug 13 00:29:21.122016 containerd[1720]: 2025-08-13 00:29:21.103 [INFO][5072] cni-plugin/k8s.go 418: Populated endpoint ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Namespace="kube-system" Pod="coredns-674b8bbfcf-hhdbb" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"2680547d-4498-4eac-9daa-64c89f065d86", ResourceVersion:"822", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"", Pod:"coredns-674b8bbfcf-hhdbb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.61.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali13161694bfa", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:21.122016 containerd[1720]: 2025-08-13 00:29:21.104 [INFO][5072] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.61.7/32] ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Namespace="kube-system" Pod="coredns-674b8bbfcf-hhdbb" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" Aug 13 00:29:21.122016 containerd[1720]: 2025-08-13 00:29:21.104 [INFO][5072] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali13161694bfa ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Namespace="kube-system" Pod="coredns-674b8bbfcf-hhdbb" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" Aug 13 00:29:21.122016 containerd[1720]: 2025-08-13 00:29:21.108 [INFO][5072] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Namespace="kube-system" Pod="coredns-674b8bbfcf-hhdbb" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" Aug 13 00:29:21.122016 containerd[1720]: 2025-08-13 00:29:21.108 [INFO][5072] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Namespace="kube-system" Pod="coredns-674b8bbfcf-hhdbb" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"2680547d-4498-4eac-9daa-64c89f065d86", ResourceVersion:"822", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b", Pod:"coredns-674b8bbfcf-hhdbb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.61.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali13161694bfa", MAC:"9a:f0:6b:91:32:5d", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:21.122016 containerd[1720]: 2025-08-13 00:29:21.119 [INFO][5072] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" Namespace="kube-system" Pod="coredns-674b8bbfcf-hhdbb" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-coredns--674b8bbfcf--hhdbb-eth0" Aug 13 00:29:21.179307 kubelet[3169]: I0813 00:29:21.179129 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-768f4c5c69-4c59z" podStartSLOduration=25.846317888 podStartE2EDuration="29.179110866s" podCreationTimestamp="2025-08-13 00:28:52 +0000 UTC" firstStartedPulling="2025-08-13 00:29:17.2536893 +0000 UTC m=+43.312720382" lastFinishedPulling="2025-08-13 00:29:20.586482277 +0000 UTC m=+46.645513360" observedRunningTime="2025-08-13 00:29:21.178589281 +0000 UTC m=+47.237620371" watchObservedRunningTime="2025-08-13 00:29:21.179110866 +0000 UTC m=+47.238141961" Aug 13 00:29:21.188713 containerd[1720]: time="2025-08-13T00:29:21.188658928Z" level=info msg="connecting to shim 57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b" address="unix:///run/containerd/s/9ae977dc48d53085c27baa7214f0e061b720aa9da3ea1a332619790948ef380b" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:29:21.214473 systemd[1]: Started cri-containerd-57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b.scope - libcontainer container 57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b. Aug 13 00:29:21.249007 containerd[1720]: time="2025-08-13T00:29:21.248982804Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\" id:\"d05e034e92b73135b14bc0c0778aaccb6f0913771b6dc3810ad1da4f3fb34581\" pid:5114 exit_status:1 exited_at:{seconds:1755044961 nanos:248556203}" Aug 13 00:29:21.260995 containerd[1720]: time="2025-08-13T00:29:21.260970721Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hhdbb,Uid:2680547d-4498-4eac-9daa-64c89f065d86,Namespace:kube-system,Attempt:0,} returns sandbox id \"57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b\"" Aug 13 00:29:21.285029 containerd[1720]: time="2025-08-13T00:29:21.284278309Z" level=info msg="CreateContainer within sandbox \"57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 00:29:21.313686 containerd[1720]: time="2025-08-13T00:29:21.313657914Z" level=info msg="Container 15fa6275df0eba23f2b23f0e23e39287d185cffd5b548fc2f9c18fb42d92369a: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:21.334841 containerd[1720]: time="2025-08-13T00:29:21.334795728Z" level=info msg="CreateContainer within sandbox \"57a288d0da967bf8b6983f21afc4d14834fa10dec5b9347875cb0e8c3030425b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"15fa6275df0eba23f2b23f0e23e39287d185cffd5b548fc2f9c18fb42d92369a\"" Aug 13 00:29:21.336147 containerd[1720]: time="2025-08-13T00:29:21.336118173Z" level=info msg="StartContainer for \"15fa6275df0eba23f2b23f0e23e39287d185cffd5b548fc2f9c18fb42d92369a\"" Aug 13 00:29:21.337314 containerd[1720]: time="2025-08-13T00:29:21.337243933Z" level=info msg="connecting to shim 15fa6275df0eba23f2b23f0e23e39287d185cffd5b548fc2f9c18fb42d92369a" address="unix:///run/containerd/s/9ae977dc48d53085c27baa7214f0e061b720aa9da3ea1a332619790948ef380b" protocol=ttrpc version=3 Aug 13 00:29:21.357601 systemd[1]: Started cri-containerd-15fa6275df0eba23f2b23f0e23e39287d185cffd5b548fc2f9c18fb42d92369a.scope - libcontainer container 15fa6275df0eba23f2b23f0e23e39287d185cffd5b548fc2f9c18fb42d92369a. Aug 13 00:29:21.389859 containerd[1720]: time="2025-08-13T00:29:21.389824451Z" level=info msg="StartContainer for \"15fa6275df0eba23f2b23f0e23e39287d185cffd5b548fc2f9c18fb42d92369a\" returns successfully" Aug 13 00:29:21.690456 systemd-networkd[1364]: calidf6d7645133: Gained IPv6LL Aug 13 00:29:22.025204 containerd[1720]: time="2025-08-13T00:29:22.025111973Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5fd87f6d8c-qtj5z,Uid:e13962cc-6aa3-411a-81be-4a870f09c685,Namespace:calico-system,Attempt:0,}" Aug 13 00:29:22.130672 systemd-networkd[1364]: calif909efe9aba: Link UP Aug 13 00:29:22.131716 systemd-networkd[1364]: calif909efe9aba: Gained carrier Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.061 [INFO][5212] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0 calico-kube-controllers-5fd87f6d8c- calico-system e13962cc-6aa3-411a-81be-4a870f09c685 824 0 2025-08-13 00:28:53 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5fd87f6d8c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4372.1.0-a-4f0cf7bbd9 calico-kube-controllers-5fd87f6d8c-qtj5z eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calif909efe9aba [] [] }} ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Namespace="calico-system" Pod="calico-kube-controllers-5fd87f6d8c-qtj5z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.061 [INFO][5212] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Namespace="calico-system" Pod="calico-kube-controllers-5fd87f6d8c-qtj5z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.080 [INFO][5225] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" HandleID="k8s-pod-network.6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.080 [INFO][5225] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" HandleID="k8s-pod-network.6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f120), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4372.1.0-a-4f0cf7bbd9", "pod":"calico-kube-controllers-5fd87f6d8c-qtj5z", "timestamp":"2025-08-13 00:29:22.080838271 +0000 UTC"}, Hostname:"ci-4372.1.0-a-4f0cf7bbd9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.081 [INFO][5225] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.081 [INFO][5225] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.081 [INFO][5225] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4372.1.0-a-4f0cf7bbd9' Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.086 [INFO][5225] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.089 [INFO][5225] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.091 [INFO][5225] ipam/ipam.go 511: Trying affinity for 192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.093 [INFO][5225] ipam/ipam.go 158: Attempting to load block cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.096 [INFO][5225] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.61.0/26 host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.096 [INFO][5225] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.61.0/26 handle="k8s-pod-network.6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.097 [INFO][5225] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.102 [INFO][5225] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.61.0/26 handle="k8s-pod-network.6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.127 [INFO][5225] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.61.8/26] block=192.168.61.0/26 handle="k8s-pod-network.6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.127 [INFO][5225] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.61.8/26] handle="k8s-pod-network.6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" host="ci-4372.1.0-a-4f0cf7bbd9" Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.127 [INFO][5225] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:29:22.148945 containerd[1720]: 2025-08-13 00:29:22.127 [INFO][5225] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.61.8/26] IPv6=[] ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" HandleID="k8s-pod-network.6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Workload="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" Aug 13 00:29:22.149451 containerd[1720]: 2025-08-13 00:29:22.128 [INFO][5212] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Namespace="calico-system" Pod="calico-kube-controllers-5fd87f6d8c-qtj5z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0", GenerateName:"calico-kube-controllers-5fd87f6d8c-", Namespace:"calico-system", SelfLink:"", UID:"e13962cc-6aa3-411a-81be-4a870f09c685", ResourceVersion:"824", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5fd87f6d8c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"", Pod:"calico-kube-controllers-5fd87f6d8c-qtj5z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.61.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calif909efe9aba", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:22.149451 containerd[1720]: 2025-08-13 00:29:22.128 [INFO][5212] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.61.8/32] ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Namespace="calico-system" Pod="calico-kube-controllers-5fd87f6d8c-qtj5z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" Aug 13 00:29:22.149451 containerd[1720]: 2025-08-13 00:29:22.128 [INFO][5212] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif909efe9aba ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Namespace="calico-system" Pod="calico-kube-controllers-5fd87f6d8c-qtj5z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" Aug 13 00:29:22.149451 containerd[1720]: 2025-08-13 00:29:22.132 [INFO][5212] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Namespace="calico-system" Pod="calico-kube-controllers-5fd87f6d8c-qtj5z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" Aug 13 00:29:22.149451 containerd[1720]: 2025-08-13 00:29:22.132 [INFO][5212] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Namespace="calico-system" Pod="calico-kube-controllers-5fd87f6d8c-qtj5z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0", GenerateName:"calico-kube-controllers-5fd87f6d8c-", Namespace:"calico-system", SelfLink:"", UID:"e13962cc-6aa3-411a-81be-4a870f09c685", ResourceVersion:"824", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 28, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5fd87f6d8c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4372.1.0-a-4f0cf7bbd9", ContainerID:"6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e", Pod:"calico-kube-controllers-5fd87f6d8c-qtj5z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.61.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calif909efe9aba", MAC:"92:cc:5c:0d:15:f8", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:29:22.149451 containerd[1720]: 2025-08-13 00:29:22.145 [INFO][5212] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" Namespace="calico-system" Pod="calico-kube-controllers-5fd87f6d8c-qtj5z" WorkloadEndpoint="ci--4372.1.0--a--4f0cf7bbd9-k8s-calico--kube--controllers--5fd87f6d8c--qtj5z-eth0" Aug 13 00:29:22.196863 kubelet[3169]: I0813 00:29:22.191960 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-hhdbb" podStartSLOduration=42.191944212 podStartE2EDuration="42.191944212s" podCreationTimestamp="2025-08-13 00:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:29:22.191550023 +0000 UTC m=+48.250581110" watchObservedRunningTime="2025-08-13 00:29:22.191944212 +0000 UTC m=+48.250975291" Aug 13 00:29:22.209065 containerd[1720]: time="2025-08-13T00:29:22.209025386Z" level=info msg="connecting to shim 6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e" address="unix:///run/containerd/s/35d373e60739a5460ede1bfec8b8adbab13fd685b087817e388d73ec71912222" namespace=k8s.io protocol=ttrpc version=3 Aug 13 00:29:22.242385 systemd[1]: Started cri-containerd-6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e.scope - libcontainer container 6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e. Aug 13 00:29:22.313591 containerd[1720]: time="2025-08-13T00:29:22.313512268Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5fd87f6d8c-qtj5z,Uid:e13962cc-6aa3-411a-81be-4a870f09c685,Namespace:calico-system,Attempt:0,} returns sandbox id \"6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e\"" Aug 13 00:29:22.342436 containerd[1720]: time="2025-08-13T00:29:22.342409928Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\" id:\"82ef569dcc8fbf9f70e5999ef5050412f1eaae71654c3a6cc324382655e8f3da\" pid:5258 exit_status:1 exited_at:{seconds:1755044962 nanos:342065043}" Aug 13 00:29:22.842470 systemd-networkd[1364]: cali13161694bfa: Gained IPv6LL Aug 13 00:29:22.899356 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1511676141.mount: Deactivated successfully. Aug 13 00:29:22.952339 containerd[1720]: time="2025-08-13T00:29:22.952302234Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:22.954395 containerd[1720]: time="2025-08-13T00:29:22.954373764Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.2: active requests=0, bytes read=33083477" Aug 13 00:29:22.957184 containerd[1720]: time="2025-08-13T00:29:22.957143224Z" level=info msg="ImageCreate event name:\"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:22.960718 containerd[1720]: time="2025-08-13T00:29:22.960676036Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:22.961306 containerd[1720]: time="2025-08-13T00:29:22.961062780Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" with image id \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\", size \"33083307\" in 2.374354942s" Aug 13 00:29:22.961306 containerd[1720]: time="2025-08-13T00:29:22.961092356Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" returns image reference \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\"" Aug 13 00:29:22.961878 containerd[1720]: time="2025-08-13T00:29:22.961853031Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Aug 13 00:29:22.966834 containerd[1720]: time="2025-08-13T00:29:22.966809969Z" level=info msg="CreateContainer within sandbox \"c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Aug 13 00:29:22.986114 containerd[1720]: time="2025-08-13T00:29:22.986077370Z" level=info msg="Container 5e9831ed2b25fd9b39ad46039487574b3045977a8b80ef5ad586e66cda6ffe89: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:23.000047 containerd[1720]: time="2025-08-13T00:29:23.000021372Z" level=info msg="CreateContainer within sandbox \"c3064841501472b1191814ea42737ce2f5a0dca5d4f98ebfb9828b4f948a0157\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"5e9831ed2b25fd9b39ad46039487574b3045977a8b80ef5ad586e66cda6ffe89\"" Aug 13 00:29:23.000561 containerd[1720]: time="2025-08-13T00:29:23.000353487Z" level=info msg="StartContainer for \"5e9831ed2b25fd9b39ad46039487574b3045977a8b80ef5ad586e66cda6ffe89\"" Aug 13 00:29:23.001793 containerd[1720]: time="2025-08-13T00:29:23.001677948Z" level=info msg="connecting to shim 5e9831ed2b25fd9b39ad46039487574b3045977a8b80ef5ad586e66cda6ffe89" address="unix:///run/containerd/s/61a8b048e40ee6205a6bb1a1c4187191c12b2213cd5cc6ff53a466b245577e93" protocol=ttrpc version=3 Aug 13 00:29:23.022421 systemd[1]: Started cri-containerd-5e9831ed2b25fd9b39ad46039487574b3045977a8b80ef5ad586e66cda6ffe89.scope - libcontainer container 5e9831ed2b25fd9b39ad46039487574b3045977a8b80ef5ad586e66cda6ffe89. Aug 13 00:29:23.061685 containerd[1720]: time="2025-08-13T00:29:23.061612811Z" level=info msg="StartContainer for \"5e9831ed2b25fd9b39ad46039487574b3045977a8b80ef5ad586e66cda6ffe89\" returns successfully" Aug 13 00:29:23.183242 kubelet[3169]: I0813 00:29:23.183139 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-59f75767c5-s7k7q" podStartSLOduration=0.978476479 podStartE2EDuration="8.183122285s" podCreationTimestamp="2025-08-13 00:29:15 +0000 UTC" firstStartedPulling="2025-08-13 00:29:15.757061934 +0000 UTC m=+41.816093029" lastFinishedPulling="2025-08-13 00:29:22.961707746 +0000 UTC m=+49.020738835" observedRunningTime="2025-08-13 00:29:23.182806238 +0000 UTC m=+49.241837327" watchObservedRunningTime="2025-08-13 00:29:23.183122285 +0000 UTC m=+49.242153371" Aug 13 00:29:24.058454 systemd-networkd[1364]: calif909efe9aba: Gained IPv6LL Aug 13 00:29:25.893197 containerd[1720]: time="2025-08-13T00:29:25.893157208Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:25.900300 containerd[1720]: time="2025-08-13T00:29:25.900256504Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=47317977" Aug 13 00:29:25.902941 containerd[1720]: time="2025-08-13T00:29:25.902884216Z" level=info msg="ImageCreate event name:\"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:25.911542 containerd[1720]: time="2025-08-13T00:29:25.911483058Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:25.912114 containerd[1720]: time="2025-08-13T00:29:25.912001705Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 2.950123411s" Aug 13 00:29:25.912114 containerd[1720]: time="2025-08-13T00:29:25.912036513Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Aug 13 00:29:25.913155 containerd[1720]: time="2025-08-13T00:29:25.913034168Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\"" Aug 13 00:29:25.918886 containerd[1720]: time="2025-08-13T00:29:25.918861651Z" level=info msg="CreateContainer within sandbox \"14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 00:29:25.937667 containerd[1720]: time="2025-08-13T00:29:25.937369311Z" level=info msg="Container 49d05d4757ca9b87658dbd5f1d5ab9c165ee86416a5f61ba0ee4fc692f18333d: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:25.952864 containerd[1720]: time="2025-08-13T00:29:25.952841476Z" level=info msg="CreateContainer within sandbox \"14639e17e974728ce69338e47d4139a0817c0de30a4d241f6aa0ffd242907544\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"49d05d4757ca9b87658dbd5f1d5ab9c165ee86416a5f61ba0ee4fc692f18333d\"" Aug 13 00:29:25.953235 containerd[1720]: time="2025-08-13T00:29:25.953169059Z" level=info msg="StartContainer for \"49d05d4757ca9b87658dbd5f1d5ab9c165ee86416a5f61ba0ee4fc692f18333d\"" Aug 13 00:29:25.954278 containerd[1720]: time="2025-08-13T00:29:25.954236587Z" level=info msg="connecting to shim 49d05d4757ca9b87658dbd5f1d5ab9c165ee86416a5f61ba0ee4fc692f18333d" address="unix:///run/containerd/s/9d0e8d62cec494e81271534a2ad51cada0189d0e9bf9f61ea6cfd941664de1b1" protocol=ttrpc version=3 Aug 13 00:29:25.974426 systemd[1]: Started cri-containerd-49d05d4757ca9b87658dbd5f1d5ab9c165ee86416a5f61ba0ee4fc692f18333d.scope - libcontainer container 49d05d4757ca9b87658dbd5f1d5ab9c165ee86416a5f61ba0ee4fc692f18333d. Aug 13 00:29:26.014043 containerd[1720]: time="2025-08-13T00:29:26.013981688Z" level=info msg="StartContainer for \"49d05d4757ca9b87658dbd5f1d5ab9c165ee86416a5f61ba0ee4fc692f18333d\" returns successfully" Aug 13 00:29:26.197798 kubelet[3169]: I0813 00:29:26.196909 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7d685458f5-fchq5" podStartSLOduration=29.52365943 podStartE2EDuration="37.196891799s" podCreationTimestamp="2025-08-13 00:28:49 +0000 UTC" firstStartedPulling="2025-08-13 00:29:18.239589978 +0000 UTC m=+44.298621066" lastFinishedPulling="2025-08-13 00:29:25.912822347 +0000 UTC m=+51.971853435" observedRunningTime="2025-08-13 00:29:26.193014221 +0000 UTC m=+52.252045319" watchObservedRunningTime="2025-08-13 00:29:26.196891799 +0000 UTC m=+52.255922884" Aug 13 00:29:27.184437 kubelet[3169]: I0813 00:29:27.183785 3169 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:29:27.472413 containerd[1720]: time="2025-08-13T00:29:27.472374208Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:27.475050 containerd[1720]: time="2025-08-13T00:29:27.475018783Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.2: active requests=0, bytes read=8759190" Aug 13 00:29:27.478425 containerd[1720]: time="2025-08-13T00:29:27.478241971Z" level=info msg="ImageCreate event name:\"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:27.483310 containerd[1720]: time="2025-08-13T00:29:27.483158599Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:27.484219 containerd[1720]: time="2025-08-13T00:29:27.483947118Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.2\" with image id \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\", size \"10251893\" in 1.570882006s" Aug 13 00:29:27.484219 containerd[1720]: time="2025-08-13T00:29:27.483982068Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\" returns image reference \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\"" Aug 13 00:29:27.485268 containerd[1720]: time="2025-08-13T00:29:27.485205590Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Aug 13 00:29:27.492007 containerd[1720]: time="2025-08-13T00:29:27.491903177Z" level=info msg="CreateContainer within sandbox \"5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Aug 13 00:29:27.518233 containerd[1720]: time="2025-08-13T00:29:27.518180652Z" level=info msg="Container bcecf5d5094501f26a17d41f046f8af1c193350e412c55e2719cb0ed664a3618: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:27.541717 containerd[1720]: time="2025-08-13T00:29:27.541671477Z" level=info msg="CreateContainer within sandbox \"5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"bcecf5d5094501f26a17d41f046f8af1c193350e412c55e2719cb0ed664a3618\"" Aug 13 00:29:27.542715 containerd[1720]: time="2025-08-13T00:29:27.542451951Z" level=info msg="StartContainer for \"bcecf5d5094501f26a17d41f046f8af1c193350e412c55e2719cb0ed664a3618\"" Aug 13 00:29:27.544389 containerd[1720]: time="2025-08-13T00:29:27.544354556Z" level=info msg="connecting to shim bcecf5d5094501f26a17d41f046f8af1c193350e412c55e2719cb0ed664a3618" address="unix:///run/containerd/s/5bb2fcb84cce5ed65c00c6b11ab1459b160b63a7349b49e235338634855ab9d9" protocol=ttrpc version=3 Aug 13 00:29:27.569712 systemd[1]: Started cri-containerd-bcecf5d5094501f26a17d41f046f8af1c193350e412c55e2719cb0ed664a3618.scope - libcontainer container bcecf5d5094501f26a17d41f046f8af1c193350e412c55e2719cb0ed664a3618. Aug 13 00:29:27.609421 containerd[1720]: time="2025-08-13T00:29:27.609393951Z" level=info msg="StartContainer for \"bcecf5d5094501f26a17d41f046f8af1c193350e412c55e2719cb0ed664a3618\" returns successfully" Aug 13 00:29:27.873329 containerd[1720]: time="2025-08-13T00:29:27.873221180Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:27.875945 containerd[1720]: time="2025-08-13T00:29:27.875909729Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=77" Aug 13 00:29:27.877320 containerd[1720]: time="2025-08-13T00:29:27.877262500Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 391.583965ms" Aug 13 00:29:27.877400 containerd[1720]: time="2025-08-13T00:29:27.877323173Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Aug 13 00:29:27.878027 containerd[1720]: time="2025-08-13T00:29:27.877950600Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\"" Aug 13 00:29:27.883838 containerd[1720]: time="2025-08-13T00:29:27.883815767Z" level=info msg="CreateContainer within sandbox \"8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 00:29:27.903143 containerd[1720]: time="2025-08-13T00:29:27.903097351Z" level=info msg="Container 47489240cd6a1cbb141794cd3347ae0b946c5ffde987809c947b173b08c0ab60: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:27.916523 containerd[1720]: time="2025-08-13T00:29:27.916501277Z" level=info msg="CreateContainer within sandbox \"8de0d84c2cf57f45ce062bc289a759d535584afa6617cfc5cdc85cde0171f16f\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"47489240cd6a1cbb141794cd3347ae0b946c5ffde987809c947b173b08c0ab60\"" Aug 13 00:29:27.916919 containerd[1720]: time="2025-08-13T00:29:27.916881168Z" level=info msg="StartContainer for \"47489240cd6a1cbb141794cd3347ae0b946c5ffde987809c947b173b08c0ab60\"" Aug 13 00:29:27.917924 containerd[1720]: time="2025-08-13T00:29:27.917899107Z" level=info msg="connecting to shim 47489240cd6a1cbb141794cd3347ae0b946c5ffde987809c947b173b08c0ab60" address="unix:///run/containerd/s/8b60eb06c6910877f06d48a17bdc5feba86c9cc888921f90d48172146eba1928" protocol=ttrpc version=3 Aug 13 00:29:27.934448 systemd[1]: Started cri-containerd-47489240cd6a1cbb141794cd3347ae0b946c5ffde987809c947b173b08c0ab60.scope - libcontainer container 47489240cd6a1cbb141794cd3347ae0b946c5ffde987809c947b173b08c0ab60. Aug 13 00:29:27.974482 containerd[1720]: time="2025-08-13T00:29:27.974457420Z" level=info msg="StartContainer for \"47489240cd6a1cbb141794cd3347ae0b946c5ffde987809c947b173b08c0ab60\" returns successfully" Aug 13 00:29:28.201529 kubelet[3169]: I0813 00:29:28.201371 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7d685458f5-h9nsv" podStartSLOduration=31.698792325 podStartE2EDuration="39.201355077s" podCreationTimestamp="2025-08-13 00:28:49 +0000 UTC" firstStartedPulling="2025-08-13 00:29:20.375263167 +0000 UTC m=+46.434294260" lastFinishedPulling="2025-08-13 00:29:27.877825924 +0000 UTC m=+53.936857012" observedRunningTime="2025-08-13 00:29:28.199382942 +0000 UTC m=+54.258414029" watchObservedRunningTime="2025-08-13 00:29:28.201355077 +0000 UTC m=+54.260386165" Aug 13 00:29:29.190401 kubelet[3169]: I0813 00:29:29.190372 3169 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:29:32.453696 containerd[1720]: time="2025-08-13T00:29:32.453654570Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:32.456364 containerd[1720]: time="2025-08-13T00:29:32.456339969Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.2: active requests=0, bytes read=51276688" Aug 13 00:29:32.458939 containerd[1720]: time="2025-08-13T00:29:32.458911760Z" level=info msg="ImageCreate event name:\"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:32.462601 containerd[1720]: time="2025-08-13T00:29:32.462555865Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:32.463025 containerd[1720]: time="2025-08-13T00:29:32.463002040Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" with image id \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\", size \"52769359\" in 4.585026305s" Aug 13 00:29:32.463078 containerd[1720]: time="2025-08-13T00:29:32.463033353Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" returns image reference \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\"" Aug 13 00:29:32.463856 containerd[1720]: time="2025-08-13T00:29:32.463820851Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\"" Aug 13 00:29:32.476956 containerd[1720]: time="2025-08-13T00:29:32.476927177Z" level=info msg="CreateContainer within sandbox \"6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Aug 13 00:29:32.494314 containerd[1720]: time="2025-08-13T00:29:32.492321818Z" level=info msg="Container 3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:32.511951 containerd[1720]: time="2025-08-13T00:29:32.511928377Z" level=info msg="CreateContainer within sandbox \"6653d7b2651b6a4a34769c899dd899cbb773ff4bd8665bd1ed38be4ea7e8b01e\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c\"" Aug 13 00:29:32.513231 containerd[1720]: time="2025-08-13T00:29:32.512271551Z" level=info msg="StartContainer for \"3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c\"" Aug 13 00:29:32.513231 containerd[1720]: time="2025-08-13T00:29:32.513156779Z" level=info msg="connecting to shim 3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c" address="unix:///run/containerd/s/35d373e60739a5460ede1bfec8b8adbab13fd685b087817e388d73ec71912222" protocol=ttrpc version=3 Aug 13 00:29:32.532416 systemd[1]: Started cri-containerd-3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c.scope - libcontainer container 3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c. Aug 13 00:29:32.571738 containerd[1720]: time="2025-08-13T00:29:32.571687147Z" level=info msg="StartContainer for \"3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c\" returns successfully" Aug 13 00:29:33.241040 containerd[1720]: time="2025-08-13T00:29:33.240984597Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c\" id:\"ca742ae0b783c93b44dcd7a053c34900ad063bca15343e38ea3dd16e65dedebe\" pid:5539 exited_at:{seconds:1755044973 nanos:240650928}" Aug 13 00:29:33.252678 kubelet[3169]: I0813 00:29:33.252620 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-5fd87f6d8c-qtj5z" podStartSLOduration=30.104761152 podStartE2EDuration="40.252604614s" podCreationTimestamp="2025-08-13 00:28:53 +0000 UTC" firstStartedPulling="2025-08-13 00:29:22.315817566 +0000 UTC m=+48.374848655" lastFinishedPulling="2025-08-13 00:29:32.463661025 +0000 UTC m=+58.522692117" observedRunningTime="2025-08-13 00:29:33.21708271 +0000 UTC m=+59.276113797" watchObservedRunningTime="2025-08-13 00:29:33.252604614 +0000 UTC m=+59.311635871" Aug 13 00:29:34.155846 containerd[1720]: time="2025-08-13T00:29:34.155257847Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:34.158879 containerd[1720]: time="2025-08-13T00:29:34.158849620Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2: active requests=0, bytes read=14703784" Aug 13 00:29:34.178113 containerd[1720]: time="2025-08-13T00:29:34.178073849Z" level=info msg="ImageCreate event name:\"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:34.182218 containerd[1720]: time="2025-08-13T00:29:34.182186516Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:29:34.182655 containerd[1720]: time="2025-08-13T00:29:34.182633891Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" with image id \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\", size \"16196439\" in 1.718786243s" Aug 13 00:29:34.182719 containerd[1720]: time="2025-08-13T00:29:34.182662199Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" returns image reference \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\"" Aug 13 00:29:34.189696 containerd[1720]: time="2025-08-13T00:29:34.189670340Z" level=info msg="CreateContainer within sandbox \"5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Aug 13 00:29:34.217600 containerd[1720]: time="2025-08-13T00:29:34.217267908Z" level=info msg="Container d20c2d5b5d05b07f9d0c511afbc3e6ce1ef02ca0a190c31725ca3cbbda53a44f: CDI devices from CRI Config.CDIDevices: []" Aug 13 00:29:34.238095 containerd[1720]: time="2025-08-13T00:29:34.238071673Z" level=info msg="CreateContainer within sandbox \"5a8f7ac2db9d5fd80ccf0f802436cba6cad49ca67815b042a41d05d37adedf91\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"d20c2d5b5d05b07f9d0c511afbc3e6ce1ef02ca0a190c31725ca3cbbda53a44f\"" Aug 13 00:29:34.238553 containerd[1720]: time="2025-08-13T00:29:34.238456323Z" level=info msg="StartContainer for \"d20c2d5b5d05b07f9d0c511afbc3e6ce1ef02ca0a190c31725ca3cbbda53a44f\"" Aug 13 00:29:34.239937 containerd[1720]: time="2025-08-13T00:29:34.239910967Z" level=info msg="connecting to shim d20c2d5b5d05b07f9d0c511afbc3e6ce1ef02ca0a190c31725ca3cbbda53a44f" address="unix:///run/containerd/s/5bb2fcb84cce5ed65c00c6b11ab1459b160b63a7349b49e235338634855ab9d9" protocol=ttrpc version=3 Aug 13 00:29:34.265427 systemd[1]: Started cri-containerd-d20c2d5b5d05b07f9d0c511afbc3e6ce1ef02ca0a190c31725ca3cbbda53a44f.scope - libcontainer container d20c2d5b5d05b07f9d0c511afbc3e6ce1ef02ca0a190c31725ca3cbbda53a44f. Aug 13 00:29:34.306474 containerd[1720]: time="2025-08-13T00:29:34.306446577Z" level=info msg="StartContainer for \"d20c2d5b5d05b07f9d0c511afbc3e6ce1ef02ca0a190c31725ca3cbbda53a44f\" returns successfully" Aug 13 00:29:35.100816 kubelet[3169]: I0813 00:29:35.100786 3169 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Aug 13 00:29:35.100816 kubelet[3169]: I0813 00:29:35.100823 3169 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Aug 13 00:29:37.173410 kubelet[3169]: I0813 00:29:37.173228 3169 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:29:37.191533 kubelet[3169]: I0813 00:29:37.190888 3169 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-ptpbv" podStartSLOduration=29.634938503 podStartE2EDuration="44.190873928s" podCreationTimestamp="2025-08-13 00:28:53 +0000 UTC" firstStartedPulling="2025-08-13 00:29:19.627321978 +0000 UTC m=+45.686353068" lastFinishedPulling="2025-08-13 00:29:34.183257398 +0000 UTC m=+60.242288493" observedRunningTime="2025-08-13 00:29:35.217740988 +0000 UTC m=+61.276772076" watchObservedRunningTime="2025-08-13 00:29:37.190873928 +0000 UTC m=+63.249905015" Aug 13 00:29:46.203099 containerd[1720]: time="2025-08-13T00:29:46.203049507Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3\" id:\"dfee52aa69d8d1fcdb070ce473bf21c55e6abadc0226c95c694d6396175a19e2\" pid:5610 exited_at:{seconds:1755044986 nanos:202778623}" Aug 13 00:29:52.317260 containerd[1720]: time="2025-08-13T00:29:52.317222111Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\" id:\"2c893f3d538176d48ddbdb679da2c195740681d46aa8440e89adf08036d099b7\" pid:5636 exited_at:{seconds:1755044992 nanos:316925825}" Aug 13 00:29:52.715562 containerd[1720]: time="2025-08-13T00:29:52.715524607Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\" id:\"9e06e227bc832283d92ee429c5c97ec3ad3a7037da35724aa923b7aa180479d4\" pid:5662 exited_at:{seconds:1755044992 nanos:715311781}" Aug 13 00:29:55.127201 kubelet[3169]: I0813 00:29:55.127164 3169 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:30:03.237882 containerd[1720]: time="2025-08-13T00:30:03.237828272Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c\" id:\"2e510df59204aec8f90edc106ddc8f1a0b9629779663d6d2805d2d747528c2a1\" pid:5696 exited_at:{seconds:1755045003 nanos:237628122}" Aug 13 00:30:16.276108 containerd[1720]: time="2025-08-13T00:30:16.276068745Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3\" id:\"1f98f24c9a83a93274ee88af39489518de10efae184734ddff8de6b7569782b0\" pid:5718 exited_at:{seconds:1755045016 nanos:275514804}" Aug 13 00:30:22.266160 containerd[1720]: time="2025-08-13T00:30:22.266118190Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\" id:\"099f240989d348385dd0d3a3a341c7da5f8bc88bb12815bb0bd35443907cf06a\" pid:5742 exited_at:{seconds:1755045022 nanos:265814681}" Aug 13 00:30:24.368510 systemd[1]: Started sshd@7-10.200.8.20:22-10.200.16.10:48150.service - OpenSSH per-connection server daemon (10.200.16.10:48150). Aug 13 00:30:25.015310 sshd[5756]: Accepted publickey for core from 10.200.16.10 port 48150 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:30:25.016620 sshd-session[5756]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:30:25.022347 systemd-logind[1708]: New session 10 of user core. Aug 13 00:30:25.034434 systemd[1]: Started session-10.scope - Session 10 of User core. Aug 13 00:30:25.517996 sshd[5758]: Connection closed by 10.200.16.10 port 48150 Aug 13 00:30:25.518523 sshd-session[5756]: pam_unix(sshd:session): session closed for user core Aug 13 00:30:25.522920 systemd[1]: sshd@7-10.200.8.20:22-10.200.16.10:48150.service: Deactivated successfully. Aug 13 00:30:25.525285 systemd[1]: session-10.scope: Deactivated successfully. Aug 13 00:30:25.526700 systemd-logind[1708]: Session 10 logged out. Waiting for processes to exit. Aug 13 00:30:25.528516 systemd-logind[1708]: Removed session 10. Aug 13 00:30:30.631544 systemd[1]: Started sshd@8-10.200.8.20:22-10.200.16.10:51490.service - OpenSSH per-connection server daemon (10.200.16.10:51490). Aug 13 00:30:31.262042 sshd[5771]: Accepted publickey for core from 10.200.16.10 port 51490 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:30:31.263113 sshd-session[5771]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:30:31.266551 systemd-logind[1708]: New session 11 of user core. Aug 13 00:30:31.271480 systemd[1]: Started session-11.scope - Session 11 of User core. Aug 13 00:30:31.764946 sshd[5773]: Connection closed by 10.200.16.10 port 51490 Aug 13 00:30:31.765414 sshd-session[5771]: pam_unix(sshd:session): session closed for user core Aug 13 00:30:31.768237 systemd[1]: sshd@8-10.200.8.20:22-10.200.16.10:51490.service: Deactivated successfully. Aug 13 00:30:31.769862 systemd[1]: session-11.scope: Deactivated successfully. Aug 13 00:30:31.770628 systemd-logind[1708]: Session 11 logged out. Waiting for processes to exit. Aug 13 00:30:31.771778 systemd-logind[1708]: Removed session 11. Aug 13 00:30:33.236232 containerd[1720]: time="2025-08-13T00:30:33.236184206Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c\" id:\"35d21f43fbc8effed75e3673aca48723ffa62002e24d1f432748b12dd00b1ebe\" pid:5797 exited_at:{seconds:1755045033 nanos:235987084}" Aug 13 00:30:36.893350 systemd[1]: Started sshd@9-10.200.8.20:22-10.200.16.10:51496.service - OpenSSH per-connection server daemon (10.200.16.10:51496). Aug 13 00:30:37.520147 sshd[5809]: Accepted publickey for core from 10.200.16.10 port 51496 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:30:37.521248 sshd-session[5809]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:30:37.525363 systemd-logind[1708]: New session 12 of user core. Aug 13 00:30:37.532459 systemd[1]: Started session-12.scope - Session 12 of User core. Aug 13 00:30:38.008896 sshd[5817]: Connection closed by 10.200.16.10 port 51496 Aug 13 00:30:38.009250 sshd-session[5809]: pam_unix(sshd:session): session closed for user core Aug 13 00:30:38.012368 systemd[1]: sshd@9-10.200.8.20:22-10.200.16.10:51496.service: Deactivated successfully. Aug 13 00:30:38.014124 systemd[1]: session-12.scope: Deactivated successfully. Aug 13 00:30:38.014875 systemd-logind[1708]: Session 12 logged out. Waiting for processes to exit. Aug 13 00:30:38.016329 systemd-logind[1708]: Removed session 12. Aug 13 00:30:38.133186 systemd[1]: Started sshd@10-10.200.8.20:22-10.200.16.10:51510.service - OpenSSH per-connection server daemon (10.200.16.10:51510). Aug 13 00:30:38.772313 sshd[5830]: Accepted publickey for core from 10.200.16.10 port 51510 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:30:38.773611 sshd-session[5830]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:30:38.777745 systemd-logind[1708]: New session 13 of user core. Aug 13 00:30:38.783457 systemd[1]: Started session-13.scope - Session 13 of User core. Aug 13 00:30:39.279424 sshd[5832]: Connection closed by 10.200.16.10 port 51510 Aug 13 00:30:39.279926 sshd-session[5830]: pam_unix(sshd:session): session closed for user core Aug 13 00:30:39.282453 systemd[1]: sshd@10-10.200.8.20:22-10.200.16.10:51510.service: Deactivated successfully. Aug 13 00:30:39.284235 systemd[1]: session-13.scope: Deactivated successfully. Aug 13 00:30:39.285933 systemd-logind[1708]: Session 13 logged out. Waiting for processes to exit. Aug 13 00:30:39.286814 systemd-logind[1708]: Removed session 13. Aug 13 00:30:39.393004 systemd[1]: Started sshd@11-10.200.8.20:22-10.200.16.10:51526.service - OpenSSH per-connection server daemon (10.200.16.10:51526). Aug 13 00:30:40.021606 sshd[5842]: Accepted publickey for core from 10.200.16.10 port 51526 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:30:40.022858 sshd-session[5842]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:30:40.027366 systemd-logind[1708]: New session 14 of user core. Aug 13 00:30:40.031497 systemd[1]: Started session-14.scope - Session 14 of User core. Aug 13 00:30:40.507882 sshd[5844]: Connection closed by 10.200.16.10 port 51526 Aug 13 00:30:40.508369 sshd-session[5842]: pam_unix(sshd:session): session closed for user core Aug 13 00:30:40.510673 systemd[1]: sshd@11-10.200.8.20:22-10.200.16.10:51526.service: Deactivated successfully. Aug 13 00:30:40.512740 systemd[1]: session-14.scope: Deactivated successfully. Aug 13 00:30:40.513712 systemd-logind[1708]: Session 14 logged out. Waiting for processes to exit. Aug 13 00:30:40.514996 systemd-logind[1708]: Removed session 14. Aug 13 00:30:40.574251 containerd[1720]: time="2025-08-13T00:30:40.574186740Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c\" id:\"07db2701842c64a9db245c5385ae4c21b41d8b59bb52f2ad4ed4156c58278385\" pid:5867 exited_at:{seconds:1755045040 nanos:573883965}" Aug 13 00:30:45.625042 systemd[1]: Started sshd@12-10.200.8.20:22-10.200.16.10:33308.service - OpenSSH per-connection server daemon (10.200.16.10:33308). Aug 13 00:30:46.200006 containerd[1720]: time="2025-08-13T00:30:46.199947311Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3\" id:\"963047f1319f39d1d701b573773324825c7f34580894e2876c77597f1971783a\" pid:5899 exit_status:1 exited_at:{seconds:1755045046 nanos:199756719}" Aug 13 00:30:46.256037 sshd[5885]: Accepted publickey for core from 10.200.16.10 port 33308 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:30:46.257010 sshd-session[5885]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:30:46.260994 systemd-logind[1708]: New session 15 of user core. Aug 13 00:30:46.270423 systemd[1]: Started session-15.scope - Session 15 of User core. Aug 13 00:30:46.752987 sshd[5910]: Connection closed by 10.200.16.10 port 33308 Aug 13 00:30:46.753469 sshd-session[5885]: pam_unix(sshd:session): session closed for user core Aug 13 00:30:46.755632 systemd[1]: sshd@12-10.200.8.20:22-10.200.16.10:33308.service: Deactivated successfully. Aug 13 00:30:46.757363 systemd[1]: session-15.scope: Deactivated successfully. Aug 13 00:30:46.758637 systemd-logind[1708]: Session 15 logged out. Waiting for processes to exit. Aug 13 00:30:46.759779 systemd-logind[1708]: Removed session 15. Aug 13 00:30:51.871418 systemd[1]: Started sshd@13-10.200.8.20:22-10.200.16.10:50864.service - OpenSSH per-connection server daemon (10.200.16.10:50864). Aug 13 00:30:52.233954 containerd[1720]: time="2025-08-13T00:30:52.233916185Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\" id:\"5760c4c9da27e6ce4a73489ec2d561d7fcef0c967e2ab02a30d79b968c14856b\" pid:5944 exited_at:{seconds:1755045052 nanos:233545401}" Aug 13 00:30:52.500461 sshd[5930]: Accepted publickey for core from 10.200.16.10 port 50864 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:30:52.501505 sshd-session[5930]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:30:52.505373 systemd-logind[1708]: New session 16 of user core. Aug 13 00:30:52.510446 systemd[1]: Started session-16.scope - Session 16 of User core. Aug 13 00:30:52.716031 containerd[1720]: time="2025-08-13T00:30:52.715990145Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\" id:\"1ebde0801c924e5874f135c922ec001e9e384928fa7732f257824edaa1387e5d\" pid:5968 exited_at:{seconds:1755045052 nanos:715785516}" Aug 13 00:30:52.990257 sshd[5954]: Connection closed by 10.200.16.10 port 50864 Aug 13 00:30:52.990720 sshd-session[5930]: pam_unix(sshd:session): session closed for user core Aug 13 00:30:52.993635 systemd[1]: sshd@13-10.200.8.20:22-10.200.16.10:50864.service: Deactivated successfully. Aug 13 00:30:52.995180 systemd[1]: session-16.scope: Deactivated successfully. Aug 13 00:30:52.995931 systemd-logind[1708]: Session 16 logged out. Waiting for processes to exit. Aug 13 00:30:52.997057 systemd-logind[1708]: Removed session 16. Aug 13 00:30:58.111574 systemd[1]: Started sshd@14-10.200.8.20:22-10.200.16.10:50880.service - OpenSSH per-connection server daemon (10.200.16.10:50880). Aug 13 00:30:58.757819 sshd[6005]: Accepted publickey for core from 10.200.16.10 port 50880 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:30:58.758872 sshd-session[6005]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:30:58.763280 systemd-logind[1708]: New session 17 of user core. Aug 13 00:30:58.767432 systemd[1]: Started session-17.scope - Session 17 of User core. Aug 13 00:30:59.257466 sshd[6007]: Connection closed by 10.200.16.10 port 50880 Aug 13 00:30:59.257944 sshd-session[6005]: pam_unix(sshd:session): session closed for user core Aug 13 00:30:59.260564 systemd[1]: sshd@14-10.200.8.20:22-10.200.16.10:50880.service: Deactivated successfully. Aug 13 00:30:59.262275 systemd[1]: session-17.scope: Deactivated successfully. Aug 13 00:30:59.263602 systemd-logind[1708]: Session 17 logged out. Waiting for processes to exit. Aug 13 00:30:59.264638 systemd-logind[1708]: Removed session 17. Aug 13 00:30:59.372769 systemd[1]: Started sshd@15-10.200.8.20:22-10.200.16.10:50882.service - OpenSSH per-connection server daemon (10.200.16.10:50882). Aug 13 00:31:00.001431 sshd[6019]: Accepted publickey for core from 10.200.16.10 port 50882 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:31:00.003102 sshd-session[6019]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:31:00.009905 systemd-logind[1708]: New session 18 of user core. Aug 13 00:31:00.018414 systemd[1]: Started session-18.scope - Session 18 of User core. Aug 13 00:31:00.566106 sshd[6021]: Connection closed by 10.200.16.10 port 50882 Aug 13 00:31:00.566615 sshd-session[6019]: pam_unix(sshd:session): session closed for user core Aug 13 00:31:00.569676 systemd[1]: sshd@15-10.200.8.20:22-10.200.16.10:50882.service: Deactivated successfully. Aug 13 00:31:00.571411 systemd[1]: session-18.scope: Deactivated successfully. Aug 13 00:31:00.572048 systemd-logind[1708]: Session 18 logged out. Waiting for processes to exit. Aug 13 00:31:00.573499 systemd-logind[1708]: Removed session 18. Aug 13 00:31:00.680974 systemd[1]: Started sshd@16-10.200.8.20:22-10.200.16.10:48932.service - OpenSSH per-connection server daemon (10.200.16.10:48932). Aug 13 00:31:01.317765 sshd[6031]: Accepted publickey for core from 10.200.16.10 port 48932 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:31:01.317584 sshd-session[6031]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:31:01.324888 systemd-logind[1708]: New session 19 of user core. Aug 13 00:31:01.331638 systemd[1]: Started session-19.scope - Session 19 of User core. Aug 13 00:31:02.301205 sshd[6033]: Connection closed by 10.200.16.10 port 48932 Aug 13 00:31:02.301706 sshd-session[6031]: pam_unix(sshd:session): session closed for user core Aug 13 00:31:02.304586 systemd[1]: sshd@16-10.200.8.20:22-10.200.16.10:48932.service: Deactivated successfully. Aug 13 00:31:02.306210 systemd[1]: session-19.scope: Deactivated successfully. Aug 13 00:31:02.307013 systemd-logind[1708]: Session 19 logged out. Waiting for processes to exit. Aug 13 00:31:02.308180 systemd-logind[1708]: Removed session 19. Aug 13 00:31:02.412200 systemd[1]: Started sshd@17-10.200.8.20:22-10.200.16.10:48938.service - OpenSSH per-connection server daemon (10.200.16.10:48938). Aug 13 00:31:03.047434 sshd[6050]: Accepted publickey for core from 10.200.16.10 port 48938 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:31:03.049013 sshd-session[6050]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:31:03.053402 systemd-logind[1708]: New session 20 of user core. Aug 13 00:31:03.060453 systemd[1]: Started session-20.scope - Session 20 of User core. Aug 13 00:31:03.281435 containerd[1720]: time="2025-08-13T00:31:03.281395881Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c\" id:\"f439b1ba3bef0b5546ffcfecb84ef1fe9d4ffec8884be7ca8567a70c26a161e1\" pid:6065 exited_at:{seconds:1755045063 nanos:280858172}" Aug 13 00:31:03.619406 sshd[6052]: Connection closed by 10.200.16.10 port 48938 Aug 13 00:31:03.619871 sshd-session[6050]: pam_unix(sshd:session): session closed for user core Aug 13 00:31:03.622530 systemd[1]: sshd@17-10.200.8.20:22-10.200.16.10:48938.service: Deactivated successfully. Aug 13 00:31:03.624274 systemd[1]: session-20.scope: Deactivated successfully. Aug 13 00:31:03.625734 systemd-logind[1708]: Session 20 logged out. Waiting for processes to exit. Aug 13 00:31:03.627449 systemd-logind[1708]: Removed session 20. Aug 13 00:31:03.736808 systemd[1]: Started sshd@18-10.200.8.20:22-10.200.16.10:48952.service - OpenSSH per-connection server daemon (10.200.16.10:48952). Aug 13 00:31:04.365884 sshd[6084]: Accepted publickey for core from 10.200.16.10 port 48952 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:31:04.366569 sshd-session[6084]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:31:04.371561 systemd-logind[1708]: New session 21 of user core. Aug 13 00:31:04.379476 systemd[1]: Started session-21.scope - Session 21 of User core. Aug 13 00:31:04.871542 sshd[6086]: Connection closed by 10.200.16.10 port 48952 Aug 13 00:31:04.873458 sshd-session[6084]: pam_unix(sshd:session): session closed for user core Aug 13 00:31:04.876931 systemd-logind[1708]: Session 21 logged out. Waiting for processes to exit. Aug 13 00:31:04.877966 systemd[1]: sshd@18-10.200.8.20:22-10.200.16.10:48952.service: Deactivated successfully. Aug 13 00:31:04.880855 systemd[1]: session-21.scope: Deactivated successfully. Aug 13 00:31:04.885563 systemd-logind[1708]: Removed session 21. Aug 13 00:31:09.994213 systemd[1]: Started sshd@19-10.200.8.20:22-10.200.16.10:48966.service - OpenSSH per-connection server daemon (10.200.16.10:48966). Aug 13 00:31:10.623847 sshd[6100]: Accepted publickey for core from 10.200.16.10 port 48966 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:31:10.627057 sshd-session[6100]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:31:10.631982 systemd-logind[1708]: New session 22 of user core. Aug 13 00:31:10.637437 systemd[1]: Started session-22.scope - Session 22 of User core. Aug 13 00:31:11.109937 sshd[6102]: Connection closed by 10.200.16.10 port 48966 Aug 13 00:31:11.110433 sshd-session[6100]: pam_unix(sshd:session): session closed for user core Aug 13 00:31:11.113473 systemd[1]: sshd@19-10.200.8.20:22-10.200.16.10:48966.service: Deactivated successfully. Aug 13 00:31:11.115181 systemd[1]: session-22.scope: Deactivated successfully. Aug 13 00:31:11.115820 systemd-logind[1708]: Session 22 logged out. Waiting for processes to exit. Aug 13 00:31:11.116912 systemd-logind[1708]: Removed session 22. Aug 13 00:31:16.201834 containerd[1720]: time="2025-08-13T00:31:16.201789077Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3bdfdae749a06682e981cd1c7199517e41c64ec22586387d21df3c36a4a430a3\" id:\"6971e52785260495c31d516e329d196be3e74bfd1041523e2152697d13b40a7f\" pid:6127 exited_at:{seconds:1755045076 nanos:201576114}" Aug 13 00:31:16.224014 systemd[1]: Started sshd@20-10.200.8.20:22-10.200.16.10:44110.service - OpenSSH per-connection server daemon (10.200.16.10:44110). Aug 13 00:31:16.851167 sshd[6141]: Accepted publickey for core from 10.200.16.10 port 44110 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:31:16.852739 sshd-session[6141]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:31:16.860341 systemd-logind[1708]: New session 23 of user core. Aug 13 00:31:16.865378 systemd[1]: Started session-23.scope - Session 23 of User core. Aug 13 00:31:17.390977 sshd[6143]: Connection closed by 10.200.16.10 port 44110 Aug 13 00:31:17.391491 sshd-session[6141]: pam_unix(sshd:session): session closed for user core Aug 13 00:31:17.395900 systemd-logind[1708]: Session 23 logged out. Waiting for processes to exit. Aug 13 00:31:17.396896 systemd[1]: sshd@20-10.200.8.20:22-10.200.16.10:44110.service: Deactivated successfully. Aug 13 00:31:17.399779 systemd[1]: session-23.scope: Deactivated successfully. Aug 13 00:31:17.402281 systemd-logind[1708]: Removed session 23. Aug 13 00:31:22.230989 containerd[1720]: time="2025-08-13T00:31:22.230947431Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ce8ed67e782059df51f923fb8066514682c841b612c70e338ae8f83a19e79943\" id:\"3b17dec89f11b5b3cd252a9aa28460553e4bfa7007a14b783f2a461d92cacc1d\" pid:6167 exited_at:{seconds:1755045082 nanos:230735069}" Aug 13 00:31:22.508419 systemd[1]: Started sshd@21-10.200.8.20:22-10.200.16.10:51124.service - OpenSSH per-connection server daemon (10.200.16.10:51124). Aug 13 00:31:23.136593 sshd[6179]: Accepted publickey for core from 10.200.16.10 port 51124 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:31:23.137723 sshd-session[6179]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:31:23.141796 systemd-logind[1708]: New session 24 of user core. Aug 13 00:31:23.148442 systemd[1]: Started session-24.scope - Session 24 of User core. Aug 13 00:31:23.630393 sshd[6181]: Connection closed by 10.200.16.10 port 51124 Aug 13 00:31:23.631548 sshd-session[6179]: pam_unix(sshd:session): session closed for user core Aug 13 00:31:23.634144 systemd[1]: sshd@21-10.200.8.20:22-10.200.16.10:51124.service: Deactivated successfully. Aug 13 00:31:23.635778 systemd[1]: session-24.scope: Deactivated successfully. Aug 13 00:31:23.637110 systemd-logind[1708]: Session 24 logged out. Waiting for processes to exit. Aug 13 00:31:23.638647 systemd-logind[1708]: Removed session 24. Aug 13 00:31:28.746511 systemd[1]: Started sshd@22-10.200.8.20:22-10.200.16.10:51136.service - OpenSSH per-connection server daemon (10.200.16.10:51136). Aug 13 00:31:29.390750 sshd[6193]: Accepted publickey for core from 10.200.16.10 port 51136 ssh2: RSA SHA256:j7p4XQXWFlakDCpIugyDQaaIj3GWiUw3GsrDiBCoheU Aug 13 00:31:29.391917 sshd-session[6193]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:31:29.396841 systemd-logind[1708]: New session 25 of user core. Aug 13 00:31:29.403460 systemd[1]: Started session-25.scope - Session 25 of User core. Aug 13 00:31:29.875256 sshd[6195]: Connection closed by 10.200.16.10 port 51136 Aug 13 00:31:29.875910 sshd-session[6193]: pam_unix(sshd:session): session closed for user core Aug 13 00:31:29.878362 systemd[1]: sshd@22-10.200.8.20:22-10.200.16.10:51136.service: Deactivated successfully. Aug 13 00:31:29.880235 systemd[1]: session-25.scope: Deactivated successfully. Aug 13 00:31:29.882050 systemd-logind[1708]: Session 25 logged out. Waiting for processes to exit. Aug 13 00:31:29.882970 systemd-logind[1708]: Removed session 25. Aug 13 00:31:33.235038 containerd[1720]: time="2025-08-13T00:31:33.234973873Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3d12b20f40dc3a9049d7e432270b05fdb634a0a8cbca095e0286b177e77b3d6c\" id:\"9f986b2ee0afd7f46daecd5ec35dad69c92c317693dd8086549bf8327353d349\" pid:6218 exited_at:{seconds:1755045093 nanos:234658922}"