Jan 21 05:49:39.829007 kernel: Linux version 6.12.66-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.1_p20250801 p4) 14.3.1 20250801, GNU ld (Gentoo 2.45 p3) 2.45.0) #1 SMP PREEMPT_DYNAMIC Wed Jan 21 03:18:28 -00 2026 Jan 21 05:49:39.829212 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=81dc9acd509cfd27a090d5b49f20e13d238e4baed94e55e81b300154aedac937 Jan 21 05:49:39.829229 kernel: BIOS-provided physical RAM map: Jan 21 05:49:39.829245 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Jan 21 05:49:39.829254 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Jan 21 05:49:39.829264 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Jan 21 05:49:39.829383 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Jan 21 05:49:39.829397 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Jan 21 05:49:39.829454 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Jan 21 05:49:39.829465 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Jan 21 05:49:39.829476 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Jan 21 05:49:39.829492 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Jan 21 05:49:39.829502 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Jan 21 05:49:39.829511 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Jan 21 05:49:39.829523 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Jan 21 05:49:39.829596 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Jan 21 05:49:39.829660 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Jan 21 05:49:39.829755 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Jan 21 05:49:39.829767 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Jan 21 05:49:39.829777 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Jan 21 05:49:39.829786 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Jan 21 05:49:39.829795 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Jan 21 05:49:39.829805 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Jan 21 05:49:39.829818 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 21 05:49:39.829828 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Jan 21 05:49:39.829837 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 21 05:49:39.829851 kernel: NX (Execute Disable) protection: active Jan 21 05:49:39.829861 kernel: APIC: Static calls initialized Jan 21 05:49:39.829872 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Jan 21 05:49:39.829883 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Jan 21 05:49:39.829893 kernel: extended physical RAM map: Jan 21 05:49:39.829903 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Jan 21 05:49:39.829913 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Jan 21 05:49:39.829922 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Jan 21 05:49:39.829933 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Jan 21 05:49:39.829942 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Jan 21 05:49:39.829953 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Jan 21 05:49:39.829967 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Jan 21 05:49:39.829977 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Jan 21 05:49:39.829987 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Jan 21 05:49:39.830002 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Jan 21 05:49:39.830016 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Jan 21 05:49:39.830026 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Jan 21 05:49:39.830037 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Jan 21 05:49:39.830049 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Jan 21 05:49:39.830060 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Jan 21 05:49:39.830071 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Jan 21 05:49:39.830083 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Jan 21 05:49:39.830095 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Jan 21 05:49:39.830105 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Jan 21 05:49:39.830120 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Jan 21 05:49:39.830131 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Jan 21 05:49:39.830142 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Jan 21 05:49:39.830153 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Jan 21 05:49:39.830164 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Jan 21 05:49:39.830175 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 21 05:49:39.830186 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Jan 21 05:49:39.830197 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 21 05:49:39.830373 kernel: efi: EFI v2.7 by EDK II Jan 21 05:49:39.830392 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Jan 21 05:49:39.830463 kernel: random: crng init done Jan 21 05:49:39.830482 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Jan 21 05:49:39.830598 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Jan 21 05:49:39.830613 kernel: secureboot: Secure boot disabled Jan 21 05:49:39.830623 kernel: SMBIOS 2.8 present. Jan 21 05:49:39.830633 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Jan 21 05:49:39.830642 kernel: DMI: Memory slots populated: 1/1 Jan 21 05:49:39.830652 kernel: Hypervisor detected: KVM Jan 21 05:49:39.830664 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Jan 21 05:49:39.830678 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jan 21 05:49:39.830688 kernel: kvm-clock: using sched offset of 38301280769 cycles Jan 21 05:49:39.830699 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jan 21 05:49:39.830715 kernel: tsc: Detected 2445.426 MHz processor Jan 21 05:49:39.830726 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 21 05:49:39.830737 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 21 05:49:39.830750 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Jan 21 05:49:39.830763 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Jan 21 05:49:39.830773 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 21 05:49:39.830784 kernel: Using GB pages for direct mapping Jan 21 05:49:39.830798 kernel: ACPI: Early table checksum verification disabled Jan 21 05:49:39.830811 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Jan 21 05:49:39.830825 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Jan 21 05:49:39.830837 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jan 21 05:49:39.830847 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 21 05:49:39.830858 kernel: ACPI: FACS 0x000000009CBDD000 000040 Jan 21 05:49:39.830868 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 21 05:49:39.830886 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 21 05:49:39.830897 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 21 05:49:39.830908 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 21 05:49:39.830919 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Jan 21 05:49:39.830930 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Jan 21 05:49:39.830943 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Jan 21 05:49:39.830955 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Jan 21 05:49:39.830969 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Jan 21 05:49:39.830980 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Jan 21 05:49:39.830991 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Jan 21 05:49:39.831005 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Jan 21 05:49:39.831015 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Jan 21 05:49:39.831025 kernel: No NUMA configuration found Jan 21 05:49:39.831036 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Jan 21 05:49:39.831047 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Jan 21 05:49:39.831064 kernel: Zone ranges: Jan 21 05:49:39.831077 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 21 05:49:39.831090 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Jan 21 05:49:39.831101 kernel: Normal empty Jan 21 05:49:39.831111 kernel: Device empty Jan 21 05:49:39.831121 kernel: Movable zone start for each node Jan 21 05:49:39.831133 kernel: Early memory node ranges Jan 21 05:49:39.831150 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Jan 21 05:49:39.831225 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Jan 21 05:49:39.831238 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Jan 21 05:49:39.831249 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Jan 21 05:49:39.831260 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Jan 21 05:49:39.831393 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Jan 21 05:49:39.831409 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Jan 21 05:49:39.831429 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Jan 21 05:49:39.831490 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Jan 21 05:49:39.831506 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 21 05:49:39.831589 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Jan 21 05:49:39.831606 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Jan 21 05:49:39.831616 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 21 05:49:39.831627 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Jan 21 05:49:39.831640 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Jan 21 05:49:39.831652 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Jan 21 05:49:39.831664 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Jan 21 05:49:39.831680 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Jan 21 05:49:39.831692 kernel: ACPI: PM-Timer IO Port: 0x608 Jan 21 05:49:39.831705 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jan 21 05:49:39.831716 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jan 21 05:49:39.831731 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jan 21 05:49:39.831742 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jan 21 05:49:39.831753 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jan 21 05:49:39.831767 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jan 21 05:49:39.831780 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jan 21 05:49:39.831790 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 21 05:49:39.831801 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jan 21 05:49:39.831816 kernel: TSC deadline timer available Jan 21 05:49:39.831829 kernel: CPU topo: Max. logical packages: 1 Jan 21 05:49:39.831841 kernel: CPU topo: Max. logical dies: 1 Jan 21 05:49:39.831854 kernel: CPU topo: Max. dies per package: 1 Jan 21 05:49:39.831866 kernel: CPU topo: Max. threads per core: 1 Jan 21 05:49:39.831877 kernel: CPU topo: Num. cores per package: 4 Jan 21 05:49:39.831889 kernel: CPU topo: Num. threads per package: 4 Jan 21 05:49:39.831909 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Jan 21 05:49:39.831920 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jan 21 05:49:39.831931 kernel: kvm-guest: KVM setup pv remote TLB flush Jan 21 05:49:39.831941 kernel: kvm-guest: setup PV sched yield Jan 21 05:49:39.831952 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Jan 21 05:49:39.831965 kernel: Booting paravirtualized kernel on KVM Jan 21 05:49:39.831979 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 21 05:49:39.831990 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Jan 21 05:49:39.832006 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Jan 21 05:49:39.832016 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Jan 21 05:49:39.832031 kernel: pcpu-alloc: [0] 0 1 2 3 Jan 21 05:49:39.832041 kernel: kvm-guest: PV spinlocks enabled Jan 21 05:49:39.832052 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jan 21 05:49:39.832165 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=81dc9acd509cfd27a090d5b49f20e13d238e4baed94e55e81b300154aedac937 Jan 21 05:49:39.832184 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 21 05:49:39.832195 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 21 05:49:39.832207 kernel: Fallback order for Node 0: 0 Jan 21 05:49:39.832220 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Jan 21 05:49:39.832232 kernel: Policy zone: DMA32 Jan 21 05:49:39.832244 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 21 05:49:39.832257 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Jan 21 05:49:39.832371 kernel: ftrace: allocating 40128 entries in 157 pages Jan 21 05:49:39.832387 kernel: ftrace: allocated 157 pages with 5 groups Jan 21 05:49:39.832400 kernel: Dynamic Preempt: voluntary Jan 21 05:49:39.832413 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 21 05:49:39.832472 kernel: rcu: RCU event tracing is enabled. Jan 21 05:49:39.832491 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Jan 21 05:49:39.832502 kernel: Trampoline variant of Tasks RCU enabled. Jan 21 05:49:39.832518 kernel: Rude variant of Tasks RCU enabled. Jan 21 05:49:39.832593 kernel: Tracing variant of Tasks RCU enabled. Jan 21 05:49:39.832607 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 21 05:49:39.832618 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Jan 21 05:49:39.832681 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 21 05:49:39.832693 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 21 05:49:39.832704 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 21 05:49:39.832715 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Jan 21 05:49:39.832734 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 21 05:49:39.832747 kernel: Console: colour dummy device 80x25 Jan 21 05:49:39.832758 kernel: printk: legacy console [ttyS0] enabled Jan 21 05:49:39.832769 kernel: ACPI: Core revision 20240827 Jan 21 05:49:39.832780 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jan 21 05:49:39.832792 kernel: APIC: Switch to symmetric I/O mode setup Jan 21 05:49:39.832804 kernel: x2apic enabled Jan 21 05:49:39.832821 kernel: APIC: Switched APIC routing to: physical x2apic Jan 21 05:49:39.832833 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jan 21 05:49:39.832846 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jan 21 05:49:39.832860 kernel: kvm-guest: setup PV IPIs Jan 21 05:49:39.832871 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 21 05:49:39.832882 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Jan 21 05:49:39.832893 kernel: Calibrating delay loop (skipped) preset value.. 4890.85 BogoMIPS (lpj=2445426) Jan 21 05:49:39.832909 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jan 21 05:49:39.832921 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jan 21 05:49:39.832934 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jan 21 05:49:39.832947 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 21 05:49:39.832960 kernel: Spectre V2 : Mitigation: Retpolines Jan 21 05:49:39.832971 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Jan 21 05:49:39.832981 kernel: Speculative Store Bypass: Vulnerable Jan 21 05:49:39.832997 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jan 21 05:49:39.833012 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jan 21 05:49:39.833074 kernel: active return thunk: srso_alias_return_thunk Jan 21 05:49:39.833088 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jan 21 05:49:39.833101 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Jan 21 05:49:39.833112 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Jan 21 05:49:39.833125 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 21 05:49:39.833145 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 21 05:49:39.833156 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 21 05:49:39.833167 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 21 05:49:39.833178 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jan 21 05:49:39.833189 kernel: Freeing SMP alternatives memory: 32K Jan 21 05:49:39.833201 kernel: pid_max: default: 32768 minimum: 301 Jan 21 05:49:39.833213 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Jan 21 05:49:39.833229 kernel: landlock: Up and running. Jan 21 05:49:39.833241 kernel: SELinux: Initializing. Jan 21 05:49:39.833253 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 21 05:49:39.833265 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 21 05:49:39.833382 kernel: smpboot: CPU0: AMD EPYC 7763 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Jan 21 05:49:39.833399 kernel: Performance Events: PMU not available due to virtualization, using software events only. Jan 21 05:49:39.833411 kernel: signal: max sigframe size: 1776 Jan 21 05:49:39.833427 kernel: rcu: Hierarchical SRCU implementation. Jan 21 05:49:39.833438 kernel: rcu: Max phase no-delay instances is 400. Jan 21 05:49:39.833450 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Jan 21 05:49:39.833464 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jan 21 05:49:39.833476 kernel: smp: Bringing up secondary CPUs ... Jan 21 05:49:39.833487 kernel: smpboot: x86: Booting SMP configuration: Jan 21 05:49:39.833498 kernel: .... node #0, CPUs: #1 #2 #3 Jan 21 05:49:39.833513 kernel: smp: Brought up 1 node, 4 CPUs Jan 21 05:49:39.833585 kernel: smpboot: Total of 4 processors activated (19563.40 BogoMIPS) Jan 21 05:49:39.833599 kernel: Memory: 2439052K/2565800K available (14336K kernel code, 2445K rwdata, 31644K rodata, 15540K init, 2496K bss, 120812K reserved, 0K cma-reserved) Jan 21 05:49:39.833612 kernel: devtmpfs: initialized Jan 21 05:49:39.833624 kernel: x86/mm: Memory block size: 128MB Jan 21 05:49:39.833640 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Jan 21 05:49:39.833651 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Jan 21 05:49:39.833667 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Jan 21 05:49:39.833678 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Jan 21 05:49:39.833690 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Jan 21 05:49:39.833704 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Jan 21 05:49:39.833716 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 21 05:49:39.833727 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Jan 21 05:49:39.833738 kernel: pinctrl core: initialized pinctrl subsystem Jan 21 05:49:39.833755 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 21 05:49:39.833769 kernel: audit: initializing netlink subsys (disabled) Jan 21 05:49:39.833780 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 21 05:49:39.833791 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 21 05:49:39.833801 kernel: audit: type=2000 audit(1768974562.758:1): state=initialized audit_enabled=0 res=1 Jan 21 05:49:39.833813 kernel: cpuidle: using governor menu Jan 21 05:49:39.833828 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 21 05:49:39.833845 kernel: dca service started, version 1.12.1 Jan 21 05:49:39.833856 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Jan 21 05:49:39.833867 kernel: PCI: Using configuration type 1 for base access Jan 21 05:49:39.833879 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 21 05:49:39.833892 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 21 05:49:39.833905 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jan 21 05:49:39.833920 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 21 05:49:39.833937 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 21 05:49:39.833948 kernel: ACPI: Added _OSI(Module Device) Jan 21 05:49:39.833959 kernel: ACPI: Added _OSI(Processor Device) Jan 21 05:49:39.833969 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 21 05:49:39.833983 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 21 05:49:39.833997 kernel: ACPI: Interpreter enabled Jan 21 05:49:39.834007 kernel: ACPI: PM: (supports S0 S3 S5) Jan 21 05:49:39.834023 kernel: ACPI: Using IOAPIC for interrupt routing Jan 21 05:49:39.834034 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 21 05:49:39.834047 kernel: PCI: Using E820 reservations for host bridge windows Jan 21 05:49:39.834061 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Jan 21 05:49:39.834072 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 21 05:49:39.834748 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 21 05:49:39.835057 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Jan 21 05:49:39.835465 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Jan 21 05:49:39.835485 kernel: PCI host bridge to bus 0000:00 Jan 21 05:49:39.835846 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 21 05:49:39.836122 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jan 21 05:49:39.836503 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 21 05:49:39.837445 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Jan 21 05:49:39.837784 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Jan 21 05:49:39.838162 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Jan 21 05:49:39.838596 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 21 05:49:39.838972 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Jan 21 05:49:39.839410 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Jan 21 05:49:39.839792 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Jan 21 05:49:39.840147 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Jan 21 05:49:39.840773 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Jan 21 05:49:39.841063 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 21 05:49:39.841582 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Jan 21 05:49:39.841982 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Jan 21 05:49:39.842262 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Jan 21 05:49:39.842729 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Jan 21 05:49:39.843036 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Jan 21 05:49:39.843422 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Jan 21 05:49:39.843759 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Jan 21 05:49:39.844038 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Jan 21 05:49:39.844422 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Jan 21 05:49:39.844765 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Jan 21 05:49:39.845044 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Jan 21 05:49:39.845469 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Jan 21 05:49:39.845841 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Jan 21 05:49:39.846157 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Jan 21 05:49:39.846652 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Jan 21 05:49:39.847385 kernel: pci 0000:00:1f.0: quirk_ich7_lpc+0x0/0xc0 took 13671 usecs Jan 21 05:49:39.847767 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Jan 21 05:49:39.848886 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Jan 21 05:49:39.849189 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Jan 21 05:49:39.849679 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Jan 21 05:49:39.849968 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Jan 21 05:49:39.849986 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jan 21 05:49:39.850000 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jan 21 05:49:39.850014 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 21 05:49:39.850033 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jan 21 05:49:39.850044 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Jan 21 05:49:39.850055 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Jan 21 05:49:39.850066 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Jan 21 05:49:39.850078 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Jan 21 05:49:39.850092 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Jan 21 05:49:39.850105 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Jan 21 05:49:39.850121 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Jan 21 05:49:39.850132 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Jan 21 05:49:39.850143 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Jan 21 05:49:39.850157 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Jan 21 05:49:39.850172 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Jan 21 05:49:39.850183 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Jan 21 05:49:39.850194 kernel: iommu: Default domain type: Translated Jan 21 05:49:39.850210 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 21 05:49:39.850223 kernel: efivars: Registered efivars operations Jan 21 05:49:39.850236 kernel: PCI: Using ACPI for IRQ routing Jan 21 05:49:39.850247 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 21 05:49:39.850258 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Jan 21 05:49:39.850268 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Jan 21 05:49:39.850389 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Jan 21 05:49:39.850403 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Jan 21 05:49:39.850423 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Jan 21 05:49:39.850431 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Jan 21 05:49:39.850439 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Jan 21 05:49:39.850447 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Jan 21 05:49:39.850802 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Jan 21 05:49:39.851258 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Jan 21 05:49:39.852263 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 21 05:49:39.852401 kernel: vgaarb: loaded Jan 21 05:49:39.852415 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jan 21 05:49:39.852427 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jan 21 05:49:39.852437 kernel: clocksource: Switched to clocksource kvm-clock Jan 21 05:49:39.852448 kernel: VFS: Disk quotas dquot_6.6.0 Jan 21 05:49:39.852460 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 21 05:49:39.852480 kernel: pnp: PnP ACPI init Jan 21 05:49:39.853625 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Jan 21 05:49:39.853650 kernel: pnp: PnP ACPI: found 6 devices Jan 21 05:49:39.853663 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 21 05:49:39.853677 kernel: NET: Registered PF_INET protocol family Jan 21 05:49:39.853690 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 21 05:49:39.853702 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 21 05:49:39.853720 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 21 05:49:39.853753 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 21 05:49:39.853768 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 21 05:49:39.853780 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 21 05:49:39.853791 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 21 05:49:39.853803 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 21 05:49:39.853818 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 21 05:49:39.853829 kernel: NET: Registered PF_XDP protocol family Jan 21 05:49:39.854123 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Jan 21 05:49:39.854468 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Jan 21 05:49:39.854819 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jan 21 05:49:39.855025 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jan 21 05:49:39.855219 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jan 21 05:49:39.855518 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Jan 21 05:49:39.855798 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Jan 21 05:49:39.855994 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Jan 21 05:49:39.856013 kernel: PCI: CLS 0 bytes, default 64 Jan 21 05:49:39.856029 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Jan 21 05:49:39.856041 kernel: Initialise system trusted keyrings Jan 21 05:49:39.856053 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 21 05:49:39.856070 kernel: Key type asymmetric registered Jan 21 05:49:39.856081 kernel: Asymmetric key parser 'x509' registered Jan 21 05:49:39.856095 kernel: hrtimer: interrupt took 5044095 ns Jan 21 05:49:39.856110 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jan 21 05:49:39.856121 kernel: io scheduler mq-deadline registered Jan 21 05:49:39.856132 kernel: io scheduler kyber registered Jan 21 05:49:39.856143 kernel: io scheduler bfq registered Jan 21 05:49:39.856160 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 21 05:49:39.856174 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Jan 21 05:49:39.856187 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Jan 21 05:49:39.856206 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Jan 21 05:49:39.856221 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 21 05:49:39.856233 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 21 05:49:39.856245 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jan 21 05:49:39.856257 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 21 05:49:39.856269 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 21 05:49:39.856402 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jan 21 05:49:39.856766 kernel: rtc_cmos 00:04: RTC can wake from S4 Jan 21 05:49:39.856990 kernel: rtc_cmos 00:04: registered as rtc0 Jan 21 05:49:39.857194 kernel: rtc_cmos 00:04: setting system clock to 2026-01-21T05:49:31 UTC (1768974571) Jan 21 05:49:39.857612 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Jan 21 05:49:39.867201 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jan 21 05:49:39.867215 kernel: efifb: probing for efifb Jan 21 05:49:39.867227 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Jan 21 05:49:39.867240 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Jan 21 05:49:39.867261 kernel: efifb: scrolling: redraw Jan 21 05:49:39.867370 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Jan 21 05:49:39.867384 kernel: Console: switching to colour frame buffer device 160x50 Jan 21 05:49:39.867398 kernel: fb0: EFI VGA frame buffer device Jan 21 05:49:39.867411 kernel: pstore: Using crash dump compression: deflate Jan 21 05:49:39.867423 kernel: pstore: Registered efi_pstore as persistent store backend Jan 21 05:49:39.867434 kernel: NET: Registered PF_INET6 protocol family Jan 21 05:49:39.867451 kernel: Segment Routing with IPv6 Jan 21 05:49:39.867466 kernel: In-situ OAM (IOAM) with IPv6 Jan 21 05:49:39.867478 kernel: NET: Registered PF_PACKET protocol family Jan 21 05:49:39.867490 kernel: Key type dns_resolver registered Jan 21 05:49:39.867501 kernel: IPI shorthand broadcast: enabled Jan 21 05:49:39.867512 kernel: sched_clock: Marking stable (9895028642, 2370720811)->(13573043307, -1307293854) Jan 21 05:49:39.867583 kernel: registered taskstats version 1 Jan 21 05:49:39.867602 kernel: Loading compiled-in X.509 certificates Jan 21 05:49:39.867616 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.66-flatcar: f14d5cffa2c990093d4ef20dbfb9c251267551e1' Jan 21 05:49:39.867630 kernel: Demotion targets for Node 0: null Jan 21 05:49:39.867641 kernel: Key type .fscrypt registered Jan 21 05:49:39.867653 kernel: Key type fscrypt-provisioning registered Jan 21 05:49:39.867664 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 21 05:49:39.867675 kernel: ima: Allocated hash algorithm: sha1 Jan 21 05:49:39.867695 kernel: ima: No architecture policies found Jan 21 05:49:39.867706 kernel: clk: Disabling unused clocks Jan 21 05:49:39.867717 kernel: Freeing unused kernel image (initmem) memory: 15540K Jan 21 05:49:39.867729 kernel: Write protecting the kernel read-only data: 47104k Jan 21 05:49:39.867742 kernel: Freeing unused kernel image (rodata/data gap) memory: 1124K Jan 21 05:49:39.867757 kernel: Run /init as init process Jan 21 05:49:39.867768 kernel: with arguments: Jan 21 05:49:39.867784 kernel: /init Jan 21 05:49:39.867795 kernel: with environment: Jan 21 05:49:39.867807 kernel: HOME=/ Jan 21 05:49:39.867822 kernel: TERM=linux Jan 21 05:49:39.867833 kernel: SCSI subsystem initialized Jan 21 05:49:39.867844 kernel: libata version 3.00 loaded. Jan 21 05:49:39.868157 kernel: ahci 0000:00:1f.2: version 3.0 Jan 21 05:49:39.868181 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Jan 21 05:49:39.868628 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Jan 21 05:49:39.868915 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Jan 21 05:49:39.869393 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Jan 21 05:49:39.869787 kernel: scsi host0: ahci Jan 21 05:49:39.870102 kernel: scsi host1: ahci Jan 21 05:49:39.870502 kernel: scsi host2: ahci Jan 21 05:49:39.870848 kernel: scsi host3: ahci Jan 21 05:49:39.871132 kernel: scsi host4: ahci Jan 21 05:49:39.871646 kernel: scsi host5: ahci Jan 21 05:49:39.871667 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 26 lpm-pol 1 Jan 21 05:49:39.871679 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 26 lpm-pol 1 Jan 21 05:49:39.871702 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 26 lpm-pol 1 Jan 21 05:49:39.871714 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 26 lpm-pol 1 Jan 21 05:49:39.871725 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 26 lpm-pol 1 Jan 21 05:49:39.871737 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 26 lpm-pol 1 Jan 21 05:49:39.871748 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Jan 21 05:49:39.871762 kernel: ata2: SATA link down (SStatus 0 SControl 300) Jan 21 05:49:39.871777 kernel: ata4: SATA link down (SStatus 0 SControl 300) Jan 21 05:49:39.871796 kernel: ata5: SATA link down (SStatus 0 SControl 300) Jan 21 05:49:39.871808 kernel: ata1: SATA link down (SStatus 0 SControl 300) Jan 21 05:49:39.871819 kernel: ata6: SATA link down (SStatus 0 SControl 300) Jan 21 05:49:39.871830 kernel: ata3.00: LPM support broken, forcing max_power Jan 21 05:49:39.871842 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Jan 21 05:49:39.871856 kernel: ata3.00: applying bridge limits Jan 21 05:49:39.871870 kernel: ata3.00: LPM support broken, forcing max_power Jan 21 05:49:39.871886 kernel: ata3.00: configured for UDMA/100 Jan 21 05:49:39.872654 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Jan 21 05:49:39.872982 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Jan 21 05:49:39.873202 kernel: virtio_blk virtio1: [vda] 27000832 512-byte logical blocks (13.8 GB/12.9 GiB) Jan 21 05:49:39.873215 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 21 05:49:39.873224 kernel: GPT:16515071 != 27000831 Jan 21 05:49:39.873240 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 21 05:49:39.873248 kernel: GPT:16515071 != 27000831 Jan 21 05:49:39.873256 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 21 05:49:39.873264 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 21 05:49:39.873719 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Jan 21 05:49:39.873746 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 21 05:49:39.874063 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jan 21 05:49:39.874091 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 21 05:49:39.874106 kernel: device-mapper: uevent: version 1.0.3 Jan 21 05:49:39.874117 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Jan 21 05:49:39.874129 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Jan 21 05:49:39.874261 kernel: raid6: avx2x4 gen() 15180 MB/s Jan 21 05:49:39.874403 kernel: raid6: avx2x2 gen() 16696 MB/s Jan 21 05:49:39.874418 kernel: raid6: avx2x1 gen() 15587 MB/s Jan 21 05:49:39.874441 kernel: raid6: using algorithm avx2x2 gen() 16696 MB/s Jan 21 05:49:39.874453 kernel: raid6: .... xor() 16400 MB/s, rmw enabled Jan 21 05:49:39.874467 kernel: raid6: using avx2x2 recovery algorithm Jan 21 05:49:39.874482 kernel: xor: automatically using best checksumming function avx Jan 21 05:49:39.874494 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 21 05:49:39.874506 kernel: BTRFS: device fsid a1ceccde-d887-4c16-9a20-b31ca68e4074 devid 1 transid 34 /dev/mapper/usr (253:0) scanned by mount (182) Jan 21 05:49:39.874518 kernel: BTRFS info (device dm-0): first mount of filesystem a1ceccde-d887-4c16-9a20-b31ca68e4074 Jan 21 05:49:39.874597 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 21 05:49:39.874611 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 21 05:49:39.874624 kernel: BTRFS info (device dm-0): enabling free space tree Jan 21 05:49:39.874635 kernel: loop: module loaded Jan 21 05:49:39.874647 kernel: loop0: detected capacity change from 0 to 100552 Jan 21 05:49:39.874658 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 21 05:49:39.874674 systemd[1]: Successfully made /usr/ read-only. Jan 21 05:49:39.874697 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 21 05:49:39.874710 systemd[1]: Detected virtualization kvm. Jan 21 05:49:39.874721 systemd[1]: Detected architecture x86-64. Jan 21 05:49:39.874734 systemd[1]: Running in initrd. Jan 21 05:49:39.874746 systemd[1]: No hostname configured, using default hostname. Jan 21 05:49:39.874763 systemd[1]: Hostname set to . Jan 21 05:49:39.874776 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Jan 21 05:49:39.874788 systemd[1]: Queued start job for default target initrd.target. Jan 21 05:49:39.874801 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 21 05:49:39.874813 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 21 05:49:39.874827 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 21 05:49:39.874841 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 21 05:49:39.874858 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 21 05:49:39.874872 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 21 05:49:39.874885 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 21 05:49:39.874898 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 21 05:49:39.874911 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 21 05:49:39.874927 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Jan 21 05:49:39.874940 systemd[1]: Reached target paths.target - Path Units. Jan 21 05:49:39.874953 systemd[1]: Reached target slices.target - Slice Units. Jan 21 05:49:39.874966 systemd[1]: Reached target swap.target - Swaps. Jan 21 05:49:39.874978 systemd[1]: Reached target timers.target - Timer Units. Jan 21 05:49:39.874993 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 21 05:49:39.875007 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 21 05:49:39.875023 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 21 05:49:39.875036 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 21 05:49:39.875047 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Jan 21 05:49:39.875061 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 21 05:49:39.875073 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 21 05:49:39.875086 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 21 05:49:39.875099 systemd[1]: Reached target sockets.target - Socket Units. Jan 21 05:49:39.875116 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 21 05:49:39.875129 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 21 05:49:39.875142 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 21 05:49:39.875154 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 21 05:49:39.875171 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Jan 21 05:49:39.875184 systemd[1]: Starting systemd-fsck-usr.service... Jan 21 05:49:39.875196 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 21 05:49:39.875212 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 21 05:49:39.875226 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 21 05:49:39.875238 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 21 05:49:39.875254 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 21 05:49:39.875267 systemd[1]: Finished systemd-fsck-usr.service. Jan 21 05:49:39.875371 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 21 05:49:39.875429 systemd-journald[319]: Collecting audit messages is enabled. Jan 21 05:49:39.875463 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 21 05:49:39.875479 kernel: audit: type=1130 audit(1768974579.866:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:39.875492 systemd-journald[319]: Journal started Jan 21 05:49:39.875593 systemd-journald[319]: Runtime Journal (/run/log/journal/5f914ee96c58453eb12a1bd66c907ddc) is 6M, max 48M, 42M free. Jan 21 05:49:39.866000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:39.904502 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 21 05:49:39.913000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:39.945920 kernel: audit: type=1130 audit(1768974579.913:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:39.946266 systemd[1]: Started systemd-journald.service - Journal Service. Jan 21 05:49:39.957000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:39.970760 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 21 05:49:40.010472 kernel: audit: type=1130 audit(1768974579.957:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.081819 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 21 05:49:40.104176 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 21 05:49:40.130650 kernel: Bridge firewalling registered Jan 21 05:49:40.129744 systemd-modules-load[324]: Inserted module 'br_netfilter' Jan 21 05:49:40.159643 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 21 05:49:40.194000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.182662 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 21 05:49:40.236756 kernel: audit: type=1130 audit(1768974580.194:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.237498 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 21 05:49:40.248000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.251494 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 21 05:49:40.283214 kernel: audit: type=1130 audit(1768974580.248:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.334000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.354449 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 21 05:49:40.374649 kernel: audit: type=1130 audit(1768974580.334:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.386006 systemd-tmpfiles[343]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Jan 21 05:49:40.403240 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 21 05:49:40.486050 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 21 05:49:40.582712 kernel: audit: type=1130 audit(1768974580.510:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.510000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.654883 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 21 05:49:40.664402 dracut-cmdline[352]: dracut-109 Jan 21 05:49:40.675000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.682840 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 21 05:49:40.710105 kernel: audit: type=1130 audit(1768974580.675:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:40.710151 kernel: audit: type=1334 audit(1768974580.679:10): prog-id=6 op=LOAD Jan 21 05:49:40.679000 audit: BPF prog-id=6 op=LOAD Jan 21 05:49:40.710408 dracut-cmdline[352]: Using kernel command line parameters: SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=81dc9acd509cfd27a090d5b49f20e13d238e4baed94e55e81b300154aedac937 Jan 21 05:49:40.968053 systemd-resolved[369]: Positive Trust Anchors: Jan 21 05:49:40.970960 systemd-resolved[369]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 21 05:49:40.970968 systemd-resolved[369]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 21 05:49:40.971010 systemd-resolved[369]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 21 05:49:41.085800 kernel: audit: type=1130 audit(1768974581.065:11): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:41.065000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:41.048200 systemd-resolved[369]: Defaulting to hostname 'linux'. Jan 21 05:49:41.055634 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 21 05:49:41.066070 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 21 05:49:41.254366 kernel: Loading iSCSI transport class v2.0-870. Jan 21 05:49:41.341783 kernel: iscsi: registered transport (tcp) Jan 21 05:49:41.436023 kernel: iscsi: registered transport (qla4xxx) Jan 21 05:49:41.436679 kernel: QLogic iSCSI HBA Driver Jan 21 05:49:41.543947 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 21 05:49:41.634507 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 21 05:49:41.657000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:41.662424 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 21 05:49:41.953787 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 21 05:49:41.990000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:41.999268 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 21 05:49:42.028010 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 21 05:49:42.176432 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 21 05:49:42.186000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:42.196000 audit: BPF prog-id=7 op=LOAD Jan 21 05:49:42.198000 audit: BPF prog-id=8 op=LOAD Jan 21 05:49:42.201637 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 21 05:49:42.375211 systemd-udevd[587]: Using default interface naming scheme 'v257'. Jan 21 05:49:42.514779 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 21 05:49:42.538000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:42.562503 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 21 05:49:42.683255 dracut-pre-trigger[636]: rd.md=0: removing MD RAID activation Jan 21 05:49:42.854775 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 21 05:49:42.895000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:42.900000 audit: BPF prog-id=9 op=LOAD Jan 21 05:49:42.910250 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 21 05:49:42.972228 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 21 05:49:42.984000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:42.992226 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 21 05:49:43.189468 systemd-networkd[722]: lo: Link UP Jan 21 05:49:43.189529 systemd-networkd[722]: lo: Gained carrier Jan 21 05:49:43.242000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:43.191100 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 21 05:49:43.243508 systemd[1]: Reached target network.target - Network. Jan 21 05:49:43.378996 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 21 05:49:43.390000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:43.397262 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 21 05:49:43.653929 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jan 21 05:49:43.748769 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jan 21 05:49:44.078879 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jan 21 05:49:44.159486 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 21 05:49:44.198934 kernel: cryptd: max_cpu_qlen set to 1000 Jan 21 05:49:44.198969 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Jan 21 05:49:44.247503 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 21 05:49:44.306437 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 21 05:49:44.314453 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 21 05:49:44.351000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:44.351641 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 21 05:49:44.368450 systemd-networkd[722]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 21 05:49:44.448896 kernel: AES CTR mode by8 optimization enabled Jan 21 05:49:44.448937 disk-uuid[770]: Primary Header is updated. Jan 21 05:49:44.448937 disk-uuid[770]: Secondary Entries is updated. Jan 21 05:49:44.448937 disk-uuid[770]: Secondary Header is updated. Jan 21 05:49:44.368500 systemd-networkd[722]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 21 05:49:44.370136 systemd-networkd[722]: eth0: Link UP Jan 21 05:49:44.372096 systemd-networkd[722]: eth0: Gained carrier Jan 21 05:49:44.372114 systemd-networkd[722]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 21 05:49:44.569269 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 21 05:49:44.573430 systemd-networkd[722]: eth0: DHCPv4 address 10.0.0.61/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 21 05:49:45.199464 kernel: kauditd_printk_skb: 12 callbacks suppressed Jan 21 05:49:45.199739 kernel: audit: type=1130 audit(1768974585.170:24): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:45.170000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:45.174156 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 21 05:49:45.174562 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 21 05:49:45.302462 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 21 05:49:45.301000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:45.356535 kernel: audit: type=1131 audit(1768974585.301:25): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:45.380945 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 21 05:49:45.456089 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 21 05:49:45.483000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:45.496767 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 21 05:49:45.534786 kernel: audit: type=1130 audit(1768974585.483:26): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:45.533444 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 21 05:49:45.545698 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 21 05:49:45.565663 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 21 05:49:45.586960 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 21 05:49:45.602000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:45.640560 kernel: audit: type=1130 audit(1768974585.602:27): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:45.900994 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 21 05:49:45.932000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:45.971054 kernel: audit: type=1130 audit(1768974585.932:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:46.054176 systemd-networkd[722]: eth0: Gained IPv6LL Jan 21 05:49:46.434840 disk-uuid[771]: Warning: The kernel is still using the old partition table. Jan 21 05:49:46.434840 disk-uuid[771]: The new table will be used at the next reboot or after you Jan 21 05:49:46.434840 disk-uuid[771]: run partprobe(8) or kpartx(8) Jan 21 05:49:46.434840 disk-uuid[771]: The operation has completed successfully. Jan 21 05:49:46.498077 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 21 05:49:46.571695 kernel: audit: type=1130 audit(1768974586.513:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:46.571739 kernel: audit: type=1131 audit(1768974586.513:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:46.513000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:46.513000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:46.498725 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 21 05:49:46.540853 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 21 05:49:46.765467 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (864) Jan 21 05:49:46.802160 kernel: BTRFS info (device vda6): first mount of filesystem 7507227a-f217-4f04-b931-d1b758f0e0f0 Jan 21 05:49:46.804128 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 21 05:49:46.862503 kernel: BTRFS info (device vda6): turning on async discard Jan 21 05:49:46.863461 kernel: BTRFS info (device vda6): enabling free space tree Jan 21 05:49:46.925524 kernel: BTRFS info (device vda6): last unmount of filesystem 7507227a-f217-4f04-b931-d1b758f0e0f0 Jan 21 05:49:46.966269 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 21 05:49:46.983000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:46.991043 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 21 05:49:47.015997 kernel: audit: type=1130 audit(1768974586.983:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:49.002146 ignition[883]: Ignition 2.24.0 Jan 21 05:49:49.003261 ignition[883]: Stage: fetch-offline Jan 21 05:49:49.004175 ignition[883]: no configs at "/usr/lib/ignition/base.d" Jan 21 05:49:49.004195 ignition[883]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 21 05:49:49.005719 ignition[883]: parsed url from cmdline: "" Jan 21 05:49:49.005726 ignition[883]: no config URL provided Jan 21 05:49:49.007105 ignition[883]: reading system config file "/usr/lib/ignition/user.ign" Jan 21 05:49:49.007132 ignition[883]: no config at "/usr/lib/ignition/user.ign" Jan 21 05:49:49.010889 ignition[883]: op(1): [started] loading QEMU firmware config module Jan 21 05:49:49.010900 ignition[883]: op(1): executing: "modprobe" "qemu_fw_cfg" Jan 21 05:49:49.257803 ignition[883]: op(1): [finished] loading QEMU firmware config module Jan 21 05:49:50.045749 ignition[883]: parsing config with SHA512: b9c4e906b3fa312abe9ff4ee80ad047da6865eee39e244570c5cb0cfda4a9b1488d973b153724ea1ad1bb3b66316d68a85206e7041ab3de3d378c8ff40165b30 Jan 21 05:49:50.174904 unknown[883]: fetched base config from "system" Jan 21 05:49:50.175817 unknown[883]: fetched user config from "qemu" Jan 21 05:49:50.197373 ignition[883]: fetch-offline: fetch-offline passed Jan 21 05:49:50.203933 ignition[883]: Ignition finished successfully Jan 21 05:49:50.213383 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 21 05:49:50.276059 kernel: audit: type=1130 audit(1768974590.232:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:50.232000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:50.235001 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jan 21 05:49:50.260911 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 21 05:49:51.112120 ignition[894]: Ignition 2.24.0 Jan 21 05:49:51.112198 ignition[894]: Stage: kargs Jan 21 05:49:51.116838 ignition[894]: no configs at "/usr/lib/ignition/base.d" Jan 21 05:49:51.116854 ignition[894]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 21 05:49:51.256221 kernel: audit: type=1130 audit(1768974591.182:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:51.182000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:51.170129 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 21 05:49:51.143121 ignition[894]: kargs: kargs passed Jan 21 05:49:51.188617 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 21 05:49:51.143231 ignition[894]: Ignition finished successfully Jan 21 05:49:51.624222 ignition[901]: Ignition 2.24.0 Jan 21 05:49:51.625866 ignition[901]: Stage: disks Jan 21 05:49:51.626444 ignition[901]: no configs at "/usr/lib/ignition/base.d" Jan 21 05:49:51.626527 ignition[901]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 21 05:49:51.690612 ignition[901]: disks: disks passed Jan 21 05:49:51.693793 ignition[901]: Ignition finished successfully Jan 21 05:49:51.751941 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 21 05:49:51.801795 kernel: audit: type=1130 audit(1768974591.772:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:51.772000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:51.782254 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 21 05:49:51.828030 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 21 05:49:51.838145 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 21 05:49:51.868008 systemd[1]: Reached target sysinit.target - System Initialization. Jan 21 05:49:51.894366 systemd[1]: Reached target basic.target - Basic System. Jan 21 05:49:51.906772 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 21 05:49:52.521229 systemd-fsck[910]: ROOT: clean, 15/456736 files, 38230/456704 blocks Jan 21 05:49:52.641967 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 21 05:49:52.691726 kernel: audit: type=1130 audit(1768974592.655:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:52.655000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:52.662466 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 21 05:49:53.693853 kernel: EXT4-fs (vda9): mounted filesystem 3ff62864-5f9e-426d-9652-a1e94c623aaa r/w with ordered data mode. Quota mode: none. Jan 21 05:49:53.697642 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 21 05:49:53.710543 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 21 05:49:53.761083 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 21 05:49:53.789985 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 21 05:49:53.803139 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 21 05:49:53.803216 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 21 05:49:53.906412 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (919) Jan 21 05:49:53.803263 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 21 05:49:53.855497 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 21 05:49:53.893877 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 21 05:49:53.988492 kernel: BTRFS info (device vda6): first mount of filesystem 7507227a-f217-4f04-b931-d1b758f0e0f0 Jan 21 05:49:53.988633 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 21 05:49:54.049018 kernel: BTRFS info (device vda6): turning on async discard Jan 21 05:49:54.055867 kernel: BTRFS info (device vda6): enabling free space tree Jan 21 05:49:54.058077 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 21 05:49:55.601129 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 21 05:49:55.680000 kernel: audit: type=1130 audit(1768974595.611:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:55.611000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:55.615900 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 21 05:49:55.679074 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 21 05:49:55.799881 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 21 05:49:55.813451 kernel: BTRFS info (device vda6): last unmount of filesystem 7507227a-f217-4f04-b931-d1b758f0e0f0 Jan 21 05:49:55.951812 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 21 05:49:56.005929 kernel: audit: type=1130 audit(1768974595.964:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:55.964000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:56.152667 ignition[1017]: INFO : Ignition 2.24.0 Jan 21 05:49:56.152667 ignition[1017]: INFO : Stage: mount Jan 21 05:49:56.152667 ignition[1017]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 21 05:49:56.152667 ignition[1017]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 21 05:49:56.268033 kernel: audit: type=1130 audit(1768974596.191:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:56.191000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:49:56.268183 ignition[1017]: INFO : mount: mount passed Jan 21 05:49:56.268183 ignition[1017]: INFO : Ignition finished successfully Jan 21 05:49:56.180993 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 21 05:49:56.198992 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 21 05:49:56.626040 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 21 05:49:56.844910 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1029) Jan 21 05:49:56.877454 kernel: BTRFS info (device vda6): first mount of filesystem 7507227a-f217-4f04-b931-d1b758f0e0f0 Jan 21 05:49:56.877549 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 21 05:49:56.980041 kernel: BTRFS info (device vda6): turning on async discard Jan 21 05:49:56.980136 kernel: BTRFS info (device vda6): enabling free space tree Jan 21 05:49:56.997050 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 21 05:49:57.273387 ignition[1046]: INFO : Ignition 2.24.0 Jan 21 05:49:57.273387 ignition[1046]: INFO : Stage: files Jan 21 05:49:57.296955 ignition[1046]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 21 05:49:57.296955 ignition[1046]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 21 05:49:57.296955 ignition[1046]: DEBUG : files: compiled without relabeling support, skipping Jan 21 05:49:57.296955 ignition[1046]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 21 05:49:57.296955 ignition[1046]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 21 05:49:57.415018 ignition[1046]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 21 05:49:57.438584 ignition[1046]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 21 05:49:57.438584 ignition[1046]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 21 05:49:57.431100 unknown[1046]: wrote ssh authorized keys file for user: core Jan 21 05:49:57.473623 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 21 05:49:57.473623 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Jan 21 05:49:57.752451 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 21 05:49:59.375032 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 21 05:49:59.375032 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 21 05:49:59.442915 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Jan 21 05:50:00.207807 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 21 05:50:06.037934 kernel: clocksource: Long readout interval, skipping watchdog check: cs_nsec: 1409598362 wd_nsec: 1409597820 Jan 21 05:50:10.834204 ignition[1046]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 21 05:50:10.834204 ignition[1046]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 21 05:50:10.868780 ignition[1046]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 21 05:50:10.868780 ignition[1046]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 21 05:50:10.868780 ignition[1046]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 21 05:50:10.868780 ignition[1046]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Jan 21 05:50:10.868780 ignition[1046]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 21 05:50:10.868780 ignition[1046]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 21 05:50:10.868780 ignition[1046]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Jan 21 05:50:10.868780 ignition[1046]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Jan 21 05:50:11.370719 ignition[1046]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Jan 21 05:50:11.432974 ignition[1046]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jan 21 05:50:11.432974 ignition[1046]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Jan 21 05:50:11.432974 ignition[1046]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Jan 21 05:50:11.432974 ignition[1046]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Jan 21 05:50:11.574255 kernel: audit: type=1130 audit(1768974611.456:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:11.456000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:11.574483 ignition[1046]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 21 05:50:11.574483 ignition[1046]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 21 05:50:11.574483 ignition[1046]: INFO : files: files passed Jan 21 05:50:11.574483 ignition[1046]: INFO : Ignition finished successfully Jan 21 05:50:11.441165 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 21 05:50:11.462052 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 21 05:50:11.574407 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 21 05:50:11.659502 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 21 05:50:11.738105 kernel: audit: type=1130 audit(1768974611.680:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:11.738153 kernel: audit: type=1131 audit(1768974611.680:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:11.680000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:11.680000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:11.659799 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 21 05:50:11.805488 initrd-setup-root-after-ignition[1078]: grep: /sysroot/oem/oem-release: No such file or directory Jan 21 05:50:11.846229 initrd-setup-root-after-ignition[1080]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 21 05:50:11.846229 initrd-setup-root-after-ignition[1080]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 21 05:50:11.879693 initrd-setup-root-after-ignition[1084]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 21 05:50:11.859506 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 21 05:50:11.929000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:11.941739 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 21 05:50:12.036096 kernel: audit: type=1130 audit(1768974611.929:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:11.996033 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 21 05:50:12.476746 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 21 05:50:12.486412 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 21 05:50:12.542202 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 21 05:50:12.540000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:12.542736 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 21 05:50:12.595504 kernel: audit: type=1130 audit(1768974612.540:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:12.634780 kernel: audit: type=1131 audit(1768974612.540:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:12.540000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:12.648203 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 21 05:50:12.664830 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 21 05:50:12.796488 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 21 05:50:12.815000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:12.842777 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 21 05:50:12.859589 kernel: audit: type=1130 audit(1768974612.815:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:12.974551 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 21 05:50:12.975546 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 21 05:50:13.008547 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 21 05:50:13.052657 systemd[1]: Stopped target timers.target - Timer Units. Jan 21 05:50:13.068243 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 21 05:50:13.080000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:13.070180 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 21 05:50:13.134572 kernel: audit: type=1131 audit(1768974613.080:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:13.138964 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 21 05:50:13.143072 systemd[1]: Stopped target basic.target - Basic System. Jan 21 05:50:13.160971 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 21 05:50:13.230565 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 21 05:50:13.254717 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 21 05:50:13.291482 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Jan 21 05:50:13.301941 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 21 05:50:13.309663 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 21 05:50:13.356673 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 21 05:50:13.375521 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 21 05:50:13.411172 systemd[1]: Stopped target swap.target - Swaps. Jan 21 05:50:13.431725 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 21 05:50:13.432958 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 21 05:50:13.494000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:13.496084 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 21 05:50:13.530015 kernel: audit: type=1131 audit(1768974613.494:47): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:13.551829 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 21 05:50:13.566765 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 21 05:50:13.571023 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 21 05:50:13.708478 kernel: audit: type=1131 audit(1768974613.653:48): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:13.653000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:13.612126 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 21 05:50:13.614187 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 21 05:50:13.728138 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 21 05:50:13.747253 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 21 05:50:13.815000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:13.818712 systemd[1]: Stopped target paths.target - Path Units. Jan 21 05:50:13.838686 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 21 05:50:13.842770 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 21 05:50:13.892783 systemd[1]: Stopped target slices.target - Slice Units. Jan 21 05:50:13.910553 systemd[1]: Stopped target sockets.target - Socket Units. Jan 21 05:50:13.950993 systemd[1]: iscsid.socket: Deactivated successfully. Jan 21 05:50:14.011000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.028000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:13.951535 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 21 05:50:14.067000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:13.987636 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 21 05:50:13.987787 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 21 05:50:13.998595 systemd[1]: systemd-journald-audit.socket: Deactivated successfully. Jan 21 05:50:13.998839 systemd[1]: Closed systemd-journald-audit.socket - Journal Audit Socket. Jan 21 05:50:14.003051 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 21 05:50:14.003526 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 21 05:50:14.170000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.012013 systemd[1]: ignition-files.service: Deactivated successfully. Jan 21 05:50:14.205000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.012223 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 21 05:50:14.033579 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 21 05:50:14.051600 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 21 05:50:14.258000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.052493 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 21 05:50:14.083137 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 21 05:50:14.101992 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 21 05:50:14.103954 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 21 05:50:14.171871 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 21 05:50:14.172203 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 21 05:50:14.206220 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 21 05:50:14.385000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.385000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.206660 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 21 05:50:14.355145 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 21 05:50:14.355732 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 21 05:50:14.457018 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 21 05:50:14.479000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.468643 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 21 05:50:14.468820 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 21 05:50:14.567206 ignition[1104]: INFO : Ignition 2.24.0 Jan 21 05:50:14.567206 ignition[1104]: INFO : Stage: umount Jan 21 05:50:14.567206 ignition[1104]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 21 05:50:14.567206 ignition[1104]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 21 05:50:14.567206 ignition[1104]: INFO : umount: umount passed Jan 21 05:50:14.567206 ignition[1104]: INFO : Ignition finished successfully Jan 21 05:50:14.590000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.668000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.575247 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 21 05:50:14.748000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.783000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.578083 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 21 05:50:14.594099 systemd[1]: Stopped target network.target - Network. Jan 21 05:50:14.849000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.633076 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 21 05:50:14.633471 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 21 05:50:14.905000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:14.668723 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 21 05:50:14.688377 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 21 05:50:14.752957 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 21 05:50:14.753542 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 21 05:50:14.784792 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 21 05:50:14.785120 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 21 05:50:14.864398 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 21 05:50:14.864705 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 21 05:50:14.932826 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 21 05:50:14.969429 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 21 05:50:15.284098 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 21 05:50:15.287130 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 21 05:50:15.400000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.443478 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 21 05:50:15.443795 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 21 05:50:15.468000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.493000 audit: BPF prog-id=6 op=UNLOAD Jan 21 05:50:15.493740 systemd[1]: Stopped target network-pre.target - Preparation for Network. Jan 21 05:50:15.498539 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 21 05:50:15.532000 audit: BPF prog-id=9 op=UNLOAD Jan 21 05:50:15.498652 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 21 05:50:15.562836 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 21 05:50:15.604261 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 21 05:50:15.605137 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 21 05:50:15.638000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.658246 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 21 05:50:15.658529 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 21 05:50:15.696234 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 21 05:50:15.692000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.696650 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 21 05:50:15.734000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.747525 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 21 05:50:15.805123 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 21 05:50:15.811876 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 21 05:50:15.858000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.859673 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 21 05:50:15.860000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.862000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.862000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.887000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.894000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.894000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.859764 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 21 05:50:15.860102 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 21 05:50:15.860161 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 21 05:50:15.860244 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 21 05:50:16.044000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:15.860500 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 21 05:50:15.862547 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 21 05:50:15.862683 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 21 05:50:15.863006 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 21 05:50:15.863096 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 21 05:50:15.869135 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 21 05:50:15.873387 systemd[1]: systemd-network-generator.service: Deactivated successfully. Jan 21 05:50:15.873483 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Jan 21 05:50:15.888255 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 21 05:50:15.888489 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 21 05:50:15.895011 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 21 05:50:15.895115 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 21 05:50:15.978720 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 21 05:50:16.033499 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 21 05:50:16.050234 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 21 05:50:16.050594 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 21 05:50:16.207000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:16.208000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:16.212116 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 21 05:50:16.237550 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 21 05:50:16.297494 systemd[1]: Switching root. Jan 21 05:50:16.372100 systemd-journald[319]: Journal stopped Jan 21 05:50:24.309823 systemd-journald[319]: Received SIGTERM from PID 1 (systemd). Jan 21 05:50:24.309917 kernel: SELinux: policy capability network_peer_controls=1 Jan 21 05:50:24.309946 kernel: SELinux: policy capability open_perms=1 Jan 21 05:50:24.309972 kernel: SELinux: policy capability extended_socket_class=1 Jan 21 05:50:24.310061 kernel: SELinux: policy capability always_check_network=0 Jan 21 05:50:24.310079 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 21 05:50:24.310102 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 21 05:50:24.310118 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 21 05:50:24.310149 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 21 05:50:24.310166 kernel: SELinux: policy capability userspace_initial_context=0 Jan 21 05:50:24.310182 kernel: kauditd_printk_skb: 34 callbacks suppressed Jan 21 05:50:24.310206 kernel: audit: type=1403 audit(1768974616.856:83): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 21 05:50:24.310237 systemd[1]: Successfully loaded SELinux policy in 272.758ms. Jan 21 05:50:24.310259 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 31.350ms. Jan 21 05:50:24.310410 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 21 05:50:24.310433 systemd[1]: Detected virtualization kvm. Jan 21 05:50:24.310451 systemd[1]: Detected architecture x86-64. Jan 21 05:50:24.310474 systemd[1]: Detected first boot. Jan 21 05:50:24.310491 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Jan 21 05:50:24.310509 kernel: audit: type=1334 audit(1768974617.231:84): prog-id=10 op=LOAD Jan 21 05:50:24.310530 kernel: audit: type=1334 audit(1768974617.231:85): prog-id=10 op=UNLOAD Jan 21 05:50:24.310551 kernel: audit: type=1334 audit(1768974617.231:86): prog-id=11 op=LOAD Jan 21 05:50:24.310568 kernel: audit: type=1334 audit(1768974617.231:87): prog-id=11 op=UNLOAD Jan 21 05:50:24.310585 zram_generator::config[1150]: No configuration found. Jan 21 05:50:24.310603 kernel: Guest personality initialized and is inactive Jan 21 05:50:24.310619 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Jan 21 05:50:24.310655 kernel: Initialized host personality Jan 21 05:50:24.310676 kernel: NET: Registered PF_VSOCK protocol family Jan 21 05:50:24.310694 systemd[1]: Populated /etc with preset unit settings. Jan 21 05:50:24.310722 kernel: audit: type=1334 audit(1768974621.267:88): prog-id=12 op=LOAD Jan 21 05:50:24.310747 kernel: audit: type=1334 audit(1768974621.267:89): prog-id=3 op=UNLOAD Jan 21 05:50:24.310766 kernel: audit: type=1334 audit(1768974621.267:90): prog-id=13 op=LOAD Jan 21 05:50:24.310784 kernel: audit: type=1334 audit(1768974621.267:91): prog-id=14 op=LOAD Jan 21 05:50:24.310808 kernel: audit: type=1334 audit(1768974621.267:92): prog-id=4 op=UNLOAD Jan 21 05:50:24.310826 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 21 05:50:24.310846 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 21 05:50:24.310867 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 21 05:50:24.310902 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 21 05:50:24.310926 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 21 05:50:24.310943 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 21 05:50:24.310960 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 21 05:50:24.311052 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 21 05:50:24.311076 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 21 05:50:24.311100 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 21 05:50:24.311118 systemd[1]: Created slice user.slice - User and Session Slice. Jan 21 05:50:24.311136 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 21 05:50:24.311154 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 21 05:50:24.311175 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 21 05:50:24.311193 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 21 05:50:24.311211 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 21 05:50:24.311233 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 21 05:50:24.311255 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 21 05:50:24.311552 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 21 05:50:24.311578 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 21 05:50:24.311599 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 21 05:50:24.311619 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 21 05:50:24.311636 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 21 05:50:24.311659 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 21 05:50:24.311678 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 21 05:50:24.311703 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 21 05:50:24.311720 systemd[1]: Reached target remote-veritysetup.target - Remote Verity Protected Volumes. Jan 21 05:50:24.311737 systemd[1]: Reached target slices.target - Slice Units. Jan 21 05:50:24.311754 systemd[1]: Reached target swap.target - Swaps. Jan 21 05:50:24.311773 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 21 05:50:24.311798 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 21 05:50:24.311819 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Jan 21 05:50:24.311840 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 21 05:50:24.311862 kernel: kauditd_printk_skb: 5 callbacks suppressed Jan 21 05:50:24.311882 kernel: audit: type=1335 audit(1768974622.948:98): pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Jan 21 05:50:24.311903 systemd[1]: Listening on systemd-mountfsd.socket - DDI File System Mounter Socket. Jan 21 05:50:24.311923 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 21 05:50:24.311948 systemd[1]: Listening on systemd-nsresourced.socket - Namespace Resource Manager Socket. Jan 21 05:50:24.311968 systemd[1]: Listening on systemd-oomd.socket - Userspace Out-Of-Memory (OOM) Killer Socket. Jan 21 05:50:24.315122 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 21 05:50:24.315148 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 21 05:50:24.315168 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 21 05:50:24.315189 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 21 05:50:24.315208 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 21 05:50:24.315233 systemd[1]: Mounting media.mount - External Media Directory... Jan 21 05:50:24.315256 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 21 05:50:24.315397 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 21 05:50:24.315423 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 21 05:50:24.315441 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 21 05:50:24.315459 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 21 05:50:24.315480 systemd[1]: Reached target machines.target - Containers. Jan 21 05:50:24.315502 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 21 05:50:24.315520 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 21 05:50:24.315542 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 21 05:50:24.315560 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 21 05:50:24.315578 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 21 05:50:24.315598 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 21 05:50:24.315617 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 21 05:50:24.315638 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 21 05:50:24.315658 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 21 05:50:24.315681 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 21 05:50:24.315699 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 21 05:50:24.315716 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 21 05:50:24.315736 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 21 05:50:24.315760 kernel: audit: type=1131 audit(1768974623.906:99): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.315778 systemd[1]: Stopped systemd-fsck-usr.service. Jan 21 05:50:24.315799 kernel: audit: type=1131 audit(1768974623.967:100): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.315824 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 21 05:50:24.315843 kernel: audit: type=1334 audit(1768974624.023:101): prog-id=14 op=UNLOAD Jan 21 05:50:24.315861 kernel: audit: type=1334 audit(1768974624.023:102): prog-id=13 op=UNLOAD Jan 21 05:50:24.315881 kernel: audit: type=1334 audit(1768974624.033:103): prog-id=15 op=LOAD Jan 21 05:50:24.315898 kernel: audit: type=1334 audit(1768974624.035:104): prog-id=16 op=LOAD Jan 21 05:50:24.315914 kernel: audit: type=1334 audit(1768974624.066:105): prog-id=17 op=LOAD Jan 21 05:50:24.315930 kernel: ACPI: bus type drm_connector registered Jan 21 05:50:24.315952 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 21 05:50:24.315967 kernel: fuse: init (API version 7.41) Jan 21 05:50:24.331571 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 21 05:50:24.335505 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 21 05:50:24.335579 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 21 05:50:24.336061 systemd-journald[1236]: Collecting audit messages is enabled. Jan 21 05:50:24.336124 kernel: audit: type=1305 audit(1768974624.304:106): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Jan 21 05:50:24.336150 kernel: audit: type=1300 audit(1768974624.304:106): arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7ffc76783000 a2=4000 a3=0 items=0 ppid=1 pid=1236 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:50:24.336171 systemd-journald[1236]: Journal started Jan 21 05:50:24.336205 systemd-journald[1236]: Runtime Journal (/run/log/journal/5f914ee96c58453eb12a1bd66c907ddc) is 6M, max 48M, 42M free. Jan 21 05:50:22.948000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Jan 21 05:50:23.906000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:23.967000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.023000 audit: BPF prog-id=14 op=UNLOAD Jan 21 05:50:24.023000 audit: BPF prog-id=13 op=UNLOAD Jan 21 05:50:24.033000 audit: BPF prog-id=15 op=LOAD Jan 21 05:50:24.035000 audit: BPF prog-id=16 op=LOAD Jan 21 05:50:24.066000 audit: BPF prog-id=17 op=LOAD Jan 21 05:50:24.304000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Jan 21 05:50:24.304000 audit[1236]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7ffc76783000 a2=4000 a3=0 items=0 ppid=1 pid=1236 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:50:24.304000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Jan 21 05:50:21.090043 systemd[1]: Queued start job for default target multi-user.target. Jan 21 05:50:21.270760 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jan 21 05:50:21.362755 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 21 05:50:21.370747 systemd[1]: systemd-journald.service: Consumed 2.180s CPU time. Jan 21 05:50:24.387600 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Jan 21 05:50:24.437912 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 21 05:50:24.467402 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 21 05:50:24.484659 systemd[1]: Started systemd-journald.service - Journal Service. Jan 21 05:50:24.494000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.499760 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 21 05:50:24.510769 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 21 05:50:24.538965 systemd[1]: Mounted media.mount - External Media Directory. Jan 21 05:50:24.551555 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 21 05:50:24.562226 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 21 05:50:24.573449 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 21 05:50:24.584173 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 21 05:50:24.610532 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 21 05:50:24.608000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.645000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.646395 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 21 05:50:24.646867 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 21 05:50:24.661659 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 21 05:50:24.662198 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 21 05:50:24.660000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.660000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.670000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.670000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.670968 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 21 05:50:24.671521 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 21 05:50:24.679000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.679000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.680891 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 21 05:50:24.681526 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 21 05:50:24.696000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.696000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.697753 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 21 05:50:24.698097 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 21 05:50:24.705912 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 21 05:50:24.706824 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 21 05:50:24.704000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.704000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.719853 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 21 05:50:24.716000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.716000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.743000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.745229 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 21 05:50:24.763000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.766758 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 21 05:50:24.779000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.782624 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Jan 21 05:50:24.809000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-load-credentials comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:24.866055 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 21 05:50:24.879704 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Jan 21 05:50:24.899624 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 21 05:50:24.942766 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 21 05:50:24.953406 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 21 05:50:24.953505 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 21 05:50:24.967761 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Jan 21 05:50:24.987729 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 21 05:50:24.987937 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 21 05:50:25.037400 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 21 05:50:25.149674 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 21 05:50:25.158851 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 21 05:50:25.174629 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 21 05:50:25.196950 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 21 05:50:25.199747 systemd-journald[1236]: Time spent on flushing to /var/log/journal/5f914ee96c58453eb12a1bd66c907ddc is 149.215ms for 1219 entries. Jan 21 05:50:25.199747 systemd-journald[1236]: System Journal (/var/log/journal/5f914ee96c58453eb12a1bd66c907ddc) is 8M, max 163.5M, 155.5M free. Jan 21 05:50:25.402700 systemd-journald[1236]: Received client request to flush runtime journal. Jan 21 05:50:25.374000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:25.203382 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 21 05:50:25.256412 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 21 05:50:25.298622 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 21 05:50:25.348074 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 21 05:50:25.375965 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 21 05:50:25.405244 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 21 05:50:26.014541 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 21 05:50:26.039000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:26.051776 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 21 05:50:26.081217 kernel: loop1: detected capacity change from 0 to 229808 Jan 21 05:50:26.082000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:26.098229 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 21 05:50:26.127601 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Jan 21 05:50:26.141612 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 21 05:50:26.157000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:26.206466 kernel: loop2: detected capacity change from 0 to 50784 Jan 21 05:50:26.267084 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 21 05:50:26.275000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:26.279677 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 21 05:50:26.281947 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Jan 21 05:50:26.301000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:26.313000 audit: BPF prog-id=18 op=LOAD Jan 21 05:50:26.313000 audit: BPF prog-id=19 op=LOAD Jan 21 05:50:26.315000 audit: BPF prog-id=20 op=LOAD Jan 21 05:50:26.335602 systemd[1]: Starting systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer... Jan 21 05:50:26.365000 audit: BPF prog-id=21 op=LOAD Jan 21 05:50:26.380744 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 21 05:50:26.415461 kernel: loop3: detected capacity change from 0 to 111560 Jan 21 05:50:26.412426 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 21 05:50:26.458000 audit: BPF prog-id=22 op=LOAD Jan 21 05:50:26.459000 audit: BPF prog-id=23 op=LOAD Jan 21 05:50:26.459000 audit: BPF prog-id=24 op=LOAD Jan 21 05:50:26.462922 systemd[1]: Starting systemd-nsresourced.service - Namespace Resource Manager... Jan 21 05:50:26.481000 audit: BPF prog-id=25 op=LOAD Jan 21 05:50:26.481000 audit: BPF prog-id=26 op=LOAD Jan 21 05:50:26.481000 audit: BPF prog-id=27 op=LOAD Jan 21 05:50:26.493801 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 21 05:50:26.779832 kernel: loop4: detected capacity change from 0 to 229808 Jan 21 05:50:26.791083 systemd-tmpfiles[1291]: ACLs are not supported, ignoring. Jan 21 05:50:26.791173 systemd-tmpfiles[1291]: ACLs are not supported, ignoring. Jan 21 05:50:26.908767 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 21 05:50:26.939000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:26.947830 kernel: loop5: detected capacity change from 0 to 50784 Jan 21 05:50:26.970407 systemd-nsresourced[1292]: Not setting up BPF subsystem, as functionality has been disabled at compile time. Jan 21 05:50:27.137404 systemd[1]: Started systemd-nsresourced.service - Namespace Resource Manager. Jan 21 05:50:27.149000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:27.162444 kernel: loop6: detected capacity change from 0 to 111560 Jan 21 05:50:27.198000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:27.188136 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 21 05:50:27.200444 (sd-merge)[1296]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes.raw'. Jan 21 05:50:27.252408 (sd-merge)[1296]: Merged extensions into '/usr'. Jan 21 05:50:27.268809 systemd[1]: Reload requested from client PID 1270 ('systemd-sysext') (unit systemd-sysext.service)... Jan 21 05:50:27.268901 systemd[1]: Reloading... Jan 21 05:50:27.890677 systemd-oomd[1289]: No swap; memory pressure usage will be degraded Jan 21 05:50:27.943549 zram_generator::config[1340]: No configuration found. Jan 21 05:50:28.154189 systemd-resolved[1290]: Positive Trust Anchors: Jan 21 05:50:28.155423 systemd-resolved[1290]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 21 05:50:28.155432 systemd-resolved[1290]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 21 05:50:28.155475 systemd-resolved[1290]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 21 05:50:28.178939 systemd-resolved[1290]: Defaulting to hostname 'linux'. Jan 21 05:50:30.437866 systemd[1]: Reloading finished in 3161 ms. Jan 21 05:50:30.611607 systemd[1]: Started systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer. Jan 21 05:50:30.647000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:30.648999 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 21 05:50:30.655236 kernel: kauditd_printk_skb: 39 callbacks suppressed Jan 21 05:50:30.655421 kernel: audit: type=1130 audit(1768974630.647:145): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:30.709000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:30.715966 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 21 05:50:30.762687 kernel: audit: type=1130 audit(1768974630.709:146): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:30.776000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:30.801447 kernel: audit: type=1130 audit(1768974630.776:147): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:30.809148 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 21 05:50:30.882667 systemd[1]: Starting ensure-sysext.service... Jan 21 05:50:30.903665 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 21 05:50:30.924000 audit: BPF prog-id=28 op=LOAD Jan 21 05:50:30.944678 kernel: audit: type=1334 audit(1768974630.924:148): prog-id=28 op=LOAD Jan 21 05:50:30.945000 audit: BPF prog-id=25 op=UNLOAD Jan 21 05:50:30.966448 kernel: audit: type=1334 audit(1768974630.945:149): prog-id=25 op=UNLOAD Jan 21 05:50:30.966599 kernel: audit: type=1334 audit(1768974630.945:150): prog-id=29 op=LOAD Jan 21 05:50:30.945000 audit: BPF prog-id=29 op=LOAD Jan 21 05:50:30.975902 kernel: audit: type=1334 audit(1768974630.945:151): prog-id=30 op=LOAD Jan 21 05:50:30.945000 audit: BPF prog-id=30 op=LOAD Jan 21 05:50:30.945000 audit: BPF prog-id=26 op=UNLOAD Jan 21 05:50:30.945000 audit: BPF prog-id=27 op=UNLOAD Jan 21 05:50:31.004467 kernel: audit: type=1334 audit(1768974630.945:152): prog-id=26 op=UNLOAD Jan 21 05:50:31.004577 kernel: audit: type=1334 audit(1768974630.945:153): prog-id=27 op=UNLOAD Jan 21 05:50:31.004612 kernel: audit: type=1334 audit(1768974630.950:154): prog-id=31 op=LOAD Jan 21 05:50:30.950000 audit: BPF prog-id=31 op=LOAD Jan 21 05:50:31.005775 systemd[1]: Reload requested from client PID 1377 ('systemctl') (unit ensure-sysext.service)... Jan 21 05:50:31.005871 systemd[1]: Reloading... Jan 21 05:50:30.951000 audit: BPF prog-id=22 op=UNLOAD Jan 21 05:50:30.951000 audit: BPF prog-id=32 op=LOAD Jan 21 05:50:30.951000 audit: BPF prog-id=33 op=LOAD Jan 21 05:50:30.951000 audit: BPF prog-id=23 op=UNLOAD Jan 21 05:50:30.951000 audit: BPF prog-id=24 op=UNLOAD Jan 21 05:50:30.960000 audit: BPF prog-id=34 op=LOAD Jan 21 05:50:30.960000 audit: BPF prog-id=21 op=UNLOAD Jan 21 05:50:30.962000 audit: BPF prog-id=35 op=LOAD Jan 21 05:50:30.962000 audit: BPF prog-id=15 op=UNLOAD Jan 21 05:50:30.962000 audit: BPF prog-id=36 op=LOAD Jan 21 05:50:30.962000 audit: BPF prog-id=37 op=LOAD Jan 21 05:50:30.962000 audit: BPF prog-id=16 op=UNLOAD Jan 21 05:50:30.962000 audit: BPF prog-id=17 op=UNLOAD Jan 21 05:50:30.967000 audit: BPF prog-id=38 op=LOAD Jan 21 05:50:30.967000 audit: BPF prog-id=18 op=UNLOAD Jan 21 05:50:30.967000 audit: BPF prog-id=39 op=LOAD Jan 21 05:50:30.967000 audit: BPF prog-id=40 op=LOAD Jan 21 05:50:30.967000 audit: BPF prog-id=19 op=UNLOAD Jan 21 05:50:30.967000 audit: BPF prog-id=20 op=UNLOAD Jan 21 05:50:31.037260 systemd-tmpfiles[1378]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Jan 21 05:50:31.037505 systemd-tmpfiles[1378]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Jan 21 05:50:31.038217 systemd-tmpfiles[1378]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 21 05:50:31.081248 systemd-tmpfiles[1378]: ACLs are not supported, ignoring. Jan 21 05:50:31.081517 systemd-tmpfiles[1378]: ACLs are not supported, ignoring. Jan 21 05:50:31.130990 systemd-tmpfiles[1378]: Detected autofs mount point /boot during canonicalization of boot. Jan 21 05:50:31.131508 systemd-tmpfiles[1378]: Skipping /boot Jan 21 05:50:31.177027 systemd-tmpfiles[1378]: Detected autofs mount point /boot during canonicalization of boot. Jan 21 05:50:31.177142 systemd-tmpfiles[1378]: Skipping /boot Jan 21 05:50:31.310707 zram_generator::config[1410]: No configuration found. Jan 21 05:50:32.144627 systemd[1]: Reloading finished in 1137 ms. Jan 21 05:50:32.267576 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 21 05:50:32.286000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:32.464000 audit: BPF prog-id=41 op=LOAD Jan 21 05:50:32.464000 audit: BPF prog-id=35 op=UNLOAD Jan 21 05:50:32.464000 audit: BPF prog-id=42 op=LOAD Jan 21 05:50:32.464000 audit: BPF prog-id=43 op=LOAD Jan 21 05:50:32.464000 audit: BPF prog-id=36 op=UNLOAD Jan 21 05:50:32.464000 audit: BPF prog-id=37 op=UNLOAD Jan 21 05:50:32.472000 audit: BPF prog-id=44 op=LOAD Jan 21 05:50:32.472000 audit: BPF prog-id=38 op=UNLOAD Jan 21 05:50:32.472000 audit: BPF prog-id=45 op=LOAD Jan 21 05:50:32.472000 audit: BPF prog-id=46 op=LOAD Jan 21 05:50:32.472000 audit: BPF prog-id=39 op=UNLOAD Jan 21 05:50:32.472000 audit: BPF prog-id=40 op=UNLOAD Jan 21 05:50:32.475000 audit: BPF prog-id=47 op=LOAD Jan 21 05:50:32.475000 audit: BPF prog-id=34 op=UNLOAD Jan 21 05:50:32.476000 audit: BPF prog-id=48 op=LOAD Jan 21 05:50:32.476000 audit: BPF prog-id=28 op=UNLOAD Jan 21 05:50:32.479000 audit: BPF prog-id=49 op=LOAD Jan 21 05:50:32.496000 audit: BPF prog-id=50 op=LOAD Jan 21 05:50:32.496000 audit: BPF prog-id=29 op=UNLOAD Jan 21 05:50:32.496000 audit: BPF prog-id=30 op=UNLOAD Jan 21 05:50:32.500000 audit: BPF prog-id=51 op=LOAD Jan 21 05:50:32.500000 audit: BPF prog-id=31 op=UNLOAD Jan 21 05:50:32.501000 audit: BPF prog-id=52 op=LOAD Jan 21 05:50:32.501000 audit: BPF prog-id=53 op=LOAD Jan 21 05:50:32.501000 audit: BPF prog-id=32 op=UNLOAD Jan 21 05:50:32.501000 audit: BPF prog-id=33 op=UNLOAD Jan 21 05:50:32.528545 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 21 05:50:32.546000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:32.671225 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 21 05:50:32.682843 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 21 05:50:32.697858 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 21 05:50:32.731891 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 21 05:50:32.769781 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 21 05:50:32.796522 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 21 05:50:32.821648 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 21 05:50:32.837904 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 21 05:50:32.838554 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 21 05:50:32.842673 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 21 05:50:32.860457 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 21 05:50:32.877539 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 21 05:50:32.887000 audit: BPF prog-id=8 op=UNLOAD Jan 21 05:50:32.887000 audit: BPF prog-id=7 op=UNLOAD Jan 21 05:50:32.889000 audit: BPF prog-id=54 op=LOAD Jan 21 05:50:32.889000 audit: BPF prog-id=55 op=LOAD Jan 21 05:50:32.897922 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 21 05:50:32.929398 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 21 05:50:32.945647 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 21 05:50:32.953497 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 21 05:50:32.966671 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 21 05:50:32.989147 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 21 05:50:32.986000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:32.987000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:32.990008 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 21 05:50:33.022000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:33.022000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:33.033853 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 21 05:50:33.034000 audit[1464]: SYSTEM_BOOT pid=1464 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Jan 21 05:50:33.034756 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 21 05:50:33.054000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:33.056000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:50:33.144224 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 21 05:50:33.144686 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 21 05:50:33.149806 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 21 05:50:33.188487 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 21 05:50:33.367981 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 21 05:50:33.392247 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 21 05:50:33.396022 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 21 05:50:33.396571 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 21 05:50:33.396784 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 21 05:50:33.414717 systemd-udevd[1462]: Using default interface naming scheme 'v257'. Jan 21 05:50:33.881000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Jan 21 05:50:33.881000 audit[1483]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe9541d730 a2=420 a3=0 items=0 ppid=1449 pid=1483 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:50:33.881000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 21 05:50:33.904195 augenrules[1483]: No rules Jan 21 05:50:34.041906 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 21 05:50:34.061927 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 21 05:50:34.078557 systemd[1]: audit-rules.service: Deactivated successfully. Jan 21 05:50:34.078976 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 21 05:50:34.093812 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 21 05:50:34.118815 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 21 05:50:34.119696 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 21 05:50:34.141064 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 21 05:50:34.141739 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 21 05:50:34.152257 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 21 05:50:34.152814 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 21 05:50:34.220598 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 21 05:50:34.254946 systemd[1]: Finished ensure-sysext.service. Jan 21 05:50:34.261928 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jan 21 05:50:34.266656 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 21 05:50:34.269973 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 21 05:50:34.278761 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 21 05:50:34.283700 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 21 05:50:34.304582 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 21 05:50:34.330556 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 21 05:50:34.632033 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 21 05:50:34.644467 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 21 05:50:34.644749 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 21 05:50:34.644888 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 21 05:50:34.677077 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 21 05:50:34.705623 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 21 05:50:34.732760 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 21 05:50:34.732825 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 21 05:50:34.734781 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 21 05:50:34.736707 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 21 05:50:34.759745 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 21 05:50:34.761186 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 21 05:50:35.292180 augenrules[1516]: /sbin/augenrules: No change Jan 21 05:50:35.337736 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 21 05:50:35.340529 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 21 05:50:35.372543 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 21 05:50:35.387031 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 21 05:50:35.387649 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 21 05:50:35.405000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 21 05:50:35.405000 audit[1550]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc8c316d00 a2=420 a3=0 items=0 ppid=1516 pid=1550 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:50:35.405000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 21 05:50:35.405000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Jan 21 05:50:35.405000 audit[1550]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc8c319190 a2=420 a3=0 items=0 ppid=1516 pid=1550 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:50:35.405000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 21 05:50:35.406768 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 21 05:50:35.412845 augenrules[1550]: No rules Jan 21 05:50:35.408543 systemd[1]: audit-rules.service: Deactivated successfully. Jan 21 05:50:35.409265 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 21 05:50:35.442585 kernel: mousedev: PS/2 mouse device common for all mice Jan 21 05:50:35.730872 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jan 21 05:50:35.851644 kernel: ACPI: button: Power Button [PWRF] Jan 21 05:50:35.883567 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 21 05:50:35.897031 systemd[1]: Reached target time-set.target - System Time Set. Jan 21 05:50:35.906533 systemd-networkd[1528]: lo: Link UP Jan 21 05:50:35.906547 systemd-networkd[1528]: lo: Gained carrier Jan 21 05:50:35.910510 systemd-networkd[1528]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 21 05:50:35.910518 systemd-networkd[1528]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 21 05:50:35.915484 systemd-networkd[1528]: eth0: Link UP Jan 21 05:50:35.918772 systemd-networkd[1528]: eth0: Gained carrier Jan 21 05:50:35.918855 systemd-networkd[1528]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 21 05:50:35.920070 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 21 05:50:35.953536 systemd-networkd[1528]: eth0: DHCPv4 address 10.0.0.61/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 21 05:50:35.955691 systemd-timesyncd[1532]: Network configuration changed, trying to establish connection. Jan 21 05:50:37.387074 systemd-timesyncd[1532]: Contacted time server 10.0.0.1:123 (10.0.0.1). Jan 21 05:50:37.387138 systemd-resolved[1290]: Clock change detected. Flushing caches. Jan 21 05:50:37.387161 systemd-timesyncd[1532]: Initial clock synchronization to Wed 2026-01-21 05:50:37.386830 UTC. Jan 21 05:50:37.388340 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 21 05:50:37.398213 systemd[1]: Reached target network.target - Network. Jan 21 05:50:37.410872 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 21 05:50:37.426965 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Jan 21 05:50:37.451102 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 21 05:50:37.780163 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Jan 21 05:50:38.394268 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Jan 21 05:50:38.396114 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Jan 21 05:50:38.402565 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 21 05:50:38.544960 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Jan 21 05:50:39.000376 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 21 05:50:39.115209 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 21 05:50:39.121297 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 21 05:50:39.165016 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 21 05:50:39.190957 systemd-networkd[1528]: eth0: Gained IPv6LL Jan 21 05:50:39.204226 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 21 05:50:39.234108 systemd[1]: Reached target network-online.target - Network is Online. Jan 21 05:50:39.503864 ldconfig[1455]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 21 05:50:39.531471 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 21 05:50:39.554997 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 21 05:50:39.739405 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 21 05:50:39.756087 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 21 05:50:39.770258 systemd[1]: Reached target sysinit.target - System Initialization. Jan 21 05:50:39.779978 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 21 05:50:39.790147 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 21 05:50:39.801762 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Jan 21 05:50:39.810762 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 21 05:50:39.824203 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 21 05:50:39.839986 systemd[1]: Started systemd-sysupdate-reboot.timer - Reboot Automatically After System Update. Jan 21 05:50:39.854922 systemd[1]: Started systemd-sysupdate.timer - Automatic System Update. Jan 21 05:50:39.865454 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 21 05:50:39.882280 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 21 05:50:39.882434 systemd[1]: Reached target paths.target - Path Units. Jan 21 05:50:39.894470 systemd[1]: Reached target timers.target - Timer Units. Jan 21 05:50:39.924094 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 21 05:50:39.944253 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 21 05:50:39.969544 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Jan 21 05:50:39.986950 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Jan 21 05:50:40.002956 systemd[1]: Reached target ssh-access.target - SSH Access Available. Jan 21 05:50:40.110797 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 21 05:50:41.067434 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Jan 21 05:50:41.122567 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 21 05:50:41.169996 systemd[1]: Reached target sockets.target - Socket Units. Jan 21 05:50:41.182267 systemd[1]: Reached target basic.target - Basic System. Jan 21 05:50:41.191146 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 21 05:50:41.191192 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 21 05:50:41.202056 systemd[1]: Starting containerd.service - containerd container runtime... Jan 21 05:50:41.237046 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jan 21 05:50:41.405985 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 21 05:50:41.576908 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 21 05:50:41.763206 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 21 05:50:41.822902 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 21 05:50:41.858099 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 21 05:50:41.992625 kernel: kvm_amd: TSC scaling supported Jan 21 05:50:41.993021 kernel: kvm_amd: Nested Virtualization enabled Jan 21 05:50:41.993252 kernel: kvm_amd: Nested Paging enabled Jan 21 05:50:41.990470 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Jan 21 05:50:42.003871 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Jan 21 05:50:42.003936 kernel: kvm_amd: PMU virtualization is disabled Jan 21 05:50:42.094919 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:50:42.112976 jq[1601]: false Jan 21 05:50:42.117069 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 21 05:50:42.159120 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 21 05:50:42.268983 extend-filesystems[1602]: Found /dev/vda6 Jan 21 05:50:42.491139 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 21 05:50:42.556618 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 21 05:50:42.569268 extend-filesystems[1602]: Found /dev/vda9 Jan 21 05:50:42.661367 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 21 05:50:42.662276 oslogin_cache_refresh[1603]: Refreshing passwd entry cache Jan 21 05:50:42.673441 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Refreshing passwd entry cache Jan 21 05:50:42.690861 extend-filesystems[1602]: Checking size of /dev/vda9 Jan 21 05:50:42.704208 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Failure getting users, quitting Jan 21 05:50:42.704208 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 21 05:50:42.704208 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Refreshing group entry cache Jan 21 05:50:42.702988 oslogin_cache_refresh[1603]: Failure getting users, quitting Jan 21 05:50:42.703020 oslogin_cache_refresh[1603]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 21 05:50:42.703090 oslogin_cache_refresh[1603]: Refreshing group entry cache Jan 21 05:50:42.707185 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 21 05:50:42.723149 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 21 05:50:42.727456 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 21 05:50:42.732284 systemd[1]: Starting update-engine.service - Update Engine... Jan 21 05:50:42.788838 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Failure getting groups, quitting Jan 21 05:50:42.788838 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 21 05:50:42.789006 extend-filesystems[1602]: Resized partition /dev/vda9 Jan 21 05:50:42.764506 oslogin_cache_refresh[1603]: Failure getting groups, quitting Jan 21 05:50:42.830015 extend-filesystems[1633]: resize2fs 1.47.3 (8-Jul-2025) Jan 21 05:50:43.011155 kernel: EXT4-fs (vda9): resizing filesystem from 456704 to 1784827 blocks Jan 21 05:50:42.790891 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 21 05:50:42.765198 oslogin_cache_refresh[1603]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 21 05:50:42.810930 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 21 05:50:42.828229 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 21 05:50:42.828895 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 21 05:50:42.829468 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Jan 21 05:50:42.830114 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Jan 21 05:50:43.012090 systemd[1]: motdgen.service: Deactivated successfully. Jan 21 05:50:43.012595 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 21 05:50:43.080957 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 21 05:50:43.129727 kernel: EXT4-fs (vda9): resized filesystem to 1784827 Jan 21 05:50:43.277050 extend-filesystems[1633]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jan 21 05:50:43.277050 extend-filesystems[1633]: old_desc_blocks = 1, new_desc_blocks = 1 Jan 21 05:50:43.277050 extend-filesystems[1633]: The filesystem on /dev/vda9 is now 1784827 (4k) blocks long. Jan 21 05:50:43.574501 jq[1631]: true Jan 21 05:50:43.131612 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 21 05:50:43.576246 extend-filesystems[1602]: Resized filesystem in /dev/vda9 Jan 21 05:50:43.132313 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 21 05:50:43.336861 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 21 05:50:43.352080 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 21 05:50:43.599004 jq[1640]: true Jan 21 05:50:43.696458 systemd[1]: coreos-metadata.service: Deactivated successfully. Jan 21 05:50:43.705793 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jan 21 05:50:43.740053 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 21 05:50:43.896434 update_engine[1626]: I20260121 05:50:43.808492 1626 main.cc:92] Flatcar Update Engine starting Jan 21 05:50:43.934220 tar[1638]: linux-amd64/LICENSE Jan 21 05:50:43.966867 tar[1638]: linux-amd64/helm Jan 21 05:50:44.054360 dbus-daemon[1599]: [system] SELinux support is enabled Jan 21 05:50:44.058258 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 21 05:50:44.076017 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 21 05:50:44.076128 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 21 05:50:44.091828 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 21 05:50:44.092006 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 21 05:50:44.101516 update_engine[1626]: I20260121 05:50:44.101241 1626 update_check_scheduler.cc:74] Next update check in 9m51s Jan 21 05:50:44.117087 systemd[1]: Started update-engine.service - Update Engine. Jan 21 05:50:44.136304 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 21 05:50:44.226407 bash[1686]: Updated "/home/core/.ssh/authorized_keys" Jan 21 05:50:44.243294 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 21 05:50:44.266257 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jan 21 05:50:45.618454 systemd-logind[1620]: Watching system buttons on /dev/input/event2 (Power Button) Jan 21 05:50:45.618501 systemd-logind[1620]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 21 05:50:45.621076 systemd-logind[1620]: New seat seat0. Jan 21 05:50:45.631112 systemd[1]: Started systemd-logind.service - User Login Management. Jan 21 05:50:45.828929 locksmithd[1684]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 21 05:50:45.948433 sshd_keygen[1667]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 21 05:50:46.059615 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 21 05:50:46.075967 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 21 05:50:46.123855 kernel: EDAC MC: Ver: 3.0.0 Jan 21 05:50:46.128313 containerd[1639]: time="2026-01-21T05:50:46Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Jan 21 05:50:46.134417 containerd[1639]: time="2026-01-21T05:50:46.134372855Z" level=info msg="starting containerd" revision=fcd43222d6b07379a4be9786bda52438f0dd16a1 version=v2.1.5 Jan 21 05:50:46.148322 systemd[1]: issuegen.service: Deactivated successfully. Jan 21 05:50:46.148968 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 21 05:50:46.169296 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 21 05:50:46.196864 containerd[1639]: time="2026-01-21T05:50:46.196810135Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="11.382µs" Jan 21 05:50:46.196997 containerd[1639]: time="2026-01-21T05:50:46.196977528Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Jan 21 05:50:46.197160 containerd[1639]: time="2026-01-21T05:50:46.197137496Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Jan 21 05:50:46.197258 containerd[1639]: time="2026-01-21T05:50:46.197240728Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Jan 21 05:50:46.197523 containerd[1639]: time="2026-01-21T05:50:46.197502477Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Jan 21 05:50:46.198497 containerd[1639]: time="2026-01-21T05:50:46.198440869Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 21 05:50:46.198941 containerd[1639]: time="2026-01-21T05:50:46.198856094Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 21 05:50:46.199012 containerd[1639]: time="2026-01-21T05:50:46.198998149Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 21 05:50:46.199381 containerd[1639]: time="2026-01-21T05:50:46.199357860Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 21 05:50:46.199445 containerd[1639]: time="2026-01-21T05:50:46.199430917Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 21 05:50:46.199502 containerd[1639]: time="2026-01-21T05:50:46.199487943Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 21 05:50:46.200176 containerd[1639]: time="2026-01-21T05:50:46.199538457Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 21 05:50:46.200748 containerd[1639]: time="2026-01-21T05:50:46.200618373Z" level=info msg="skip loading plugin" error="EROFS unsupported, please `modprobe erofs`: skip plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 21 05:50:46.200825 containerd[1639]: time="2026-01-21T05:50:46.200808368Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Jan 21 05:50:46.201322 containerd[1639]: time="2026-01-21T05:50:46.201219976Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Jan 21 05:50:46.202338 containerd[1639]: time="2026-01-21T05:50:46.202313567Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 21 05:50:46.202851 containerd[1639]: time="2026-01-21T05:50:46.202829170Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 21 05:50:46.205293 containerd[1639]: time="2026-01-21T05:50:46.205269665Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Jan 21 05:50:46.209741 containerd[1639]: time="2026-01-21T05:50:46.208983266Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Jan 21 05:50:46.245524 containerd[1639]: time="2026-01-21T05:50:46.245077427Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Jan 21 05:50:46.268540 containerd[1639]: time="2026-01-21T05:50:46.268450775Z" level=info msg="metadata content store policy set" policy=shared Jan 21 05:50:46.905979 containerd[1639]: time="2026-01-21T05:50:46.905395601Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Jan 21 05:50:46.908135 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.928394249Z" level=info msg="loading plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.928874545Z" level=info msg="skip loading plugin" error="could not find mkfs.erofs: exec: \"mkfs.erofs\": executable file not found in $PATH: skip plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934303589Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934345648Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934434694Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934465120Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934486039Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934512369Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934534169Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934624568Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934857323Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934889262Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Jan 21 05:50:46.936091 containerd[1639]: time="2026-01-21T05:50:46.934912546Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.935360672Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.935399394Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.935430753Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.935916880Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.935948228Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.935973315Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.935999785Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.936069505Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.936093740Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.936114939Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.936134215Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.936178889Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.936336954Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Jan 21 05:50:46.936697 containerd[1639]: time="2026-01-21T05:50:46.936419568Z" level=info msg="Start snapshots syncer" Jan 21 05:50:46.985881 containerd[1639]: time="2026-01-21T05:50:46.946055073Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Jan 21 05:50:46.985881 containerd[1639]: time="2026-01-21T05:50:46.970926336Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"cgroupWritable\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"\",\"binDirs\":[\"/opt/cni/bin\"],\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogLineSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Jan 21 05:50:46.973284 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.972934705Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.973028981Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.976837139Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.976973383Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977002187Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977025130Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977051449Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977068150Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977090812Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977114567Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977139032Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977793835Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977830774Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 21 05:50:46.986453 containerd[1639]: time="2026-01-21T05:50:46.977850260Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.977865529Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.977884675Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.977903510Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.977926042Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.977952741Z" level=info msg="runtime interface created" Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.977962369Z" level=info msg="created NRI interface" Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.978064269Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.978087262Z" level=info msg="Connect containerd service" Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.978130343Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 21 05:50:46.990127 containerd[1639]: time="2026-01-21T05:50:46.984106517Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 21 05:50:47.004199 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 21 05:50:47.026950 systemd[1]: Reached target getty.target - Login Prompts. Jan 21 05:50:49.826893 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 21 05:50:50.018007 tar[1638]: linux-amd64/README.md Jan 21 05:50:50.084085 systemd[1]: Started sshd@0-10.0.0.61:22-10.0.0.1:53454.service - OpenSSH per-connection server daemon (10.0.0.1:53454). Jan 21 05:50:50.432204 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 21 05:50:51.467383 containerd[1639]: time="2026-01-21T05:50:51.466139390Z" level=info msg="Start subscribing containerd event" Jan 21 05:50:51.473377 containerd[1639]: time="2026-01-21T05:50:51.473112666Z" level=info msg="Start recovering state" Jan 21 05:50:51.476782 containerd[1639]: time="2026-01-21T05:50:51.476551214Z" level=info msg="Start event monitor" Jan 21 05:50:51.478049 containerd[1639]: time="2026-01-21T05:50:51.477954523Z" level=info msg="Start cni network conf syncer for default" Jan 21 05:50:51.479050 containerd[1639]: time="2026-01-21T05:50:51.479021004Z" level=info msg="Start streaming server" Jan 21 05:50:51.479294 containerd[1639]: time="2026-01-21T05:50:51.479267284Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Jan 21 05:50:51.479492 containerd[1639]: time="2026-01-21T05:50:51.479470723Z" level=info msg="runtime interface starting up..." Jan 21 05:50:51.487563 containerd[1639]: time="2026-01-21T05:50:51.482541027Z" level=info msg="starting plugins..." Jan 21 05:50:51.487563 containerd[1639]: time="2026-01-21T05:50:51.480972008Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 21 05:50:51.487563 containerd[1639]: time="2026-01-21T05:50:51.485216120Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 21 05:50:51.487563 containerd[1639]: time="2026-01-21T05:50:51.485286769Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Jan 21 05:50:51.487563 containerd[1639]: time="2026-01-21T05:50:51.486417119Z" level=info msg="containerd successfully booted in 5.359263s" Jan 21 05:50:51.491217 systemd[1]: Started containerd.service - containerd container runtime. Jan 21 05:50:51.598950 sshd[1720]: Accepted publickey for core from 10.0.0.1 port 53454 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:50:51.604442 sshd-session[1720]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:50:51.694530 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 21 05:50:51.722423 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 21 05:50:51.782792 systemd-logind[1620]: New session 1 of user core. Jan 21 05:50:51.867898 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 21 05:50:51.885556 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 21 05:50:51.995317 (systemd)[1740]: pam_unix(systemd-user:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:50:52.017053 systemd-logind[1620]: New session 2 of user core. Jan 21 05:50:53.131135 systemd[1740]: Queued start job for default target default.target. Jan 21 05:50:53.168356 systemd[1740]: Created slice app.slice - User Application Slice. Jan 21 05:50:53.168469 systemd[1740]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. Jan 21 05:50:53.168491 systemd[1740]: Reached target paths.target - Paths. Jan 21 05:50:53.169028 systemd[1740]: Reached target timers.target - Timers. Jan 21 05:50:53.181175 systemd[1740]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 21 05:50:53.189088 systemd[1740]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... Jan 21 05:50:53.261193 systemd[1740]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 21 05:50:53.261514 systemd[1740]: Reached target sockets.target - Sockets. Jan 21 05:50:53.276947 systemd[1740]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. Jan 21 05:50:53.277397 systemd[1740]: Reached target basic.target - Basic System. Jan 21 05:50:53.278951 systemd[1740]: Reached target default.target - Main User Target. Jan 21 05:50:53.279070 systemd[1740]: Startup finished in 1.077s. Jan 21 05:50:53.282236 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 21 05:50:53.299263 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 21 05:50:53.462903 systemd[1]: Started sshd@1-10.0.0.61:22-10.0.0.1:53458.service - OpenSSH per-connection server daemon (10.0.0.1:53458). Jan 21 05:50:53.829068 sshd[1758]: Accepted publickey for core from 10.0.0.1 port 53458 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:50:53.833438 sshd-session[1758]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:50:53.900556 systemd-logind[1620]: New session 3 of user core. Jan 21 05:50:53.937163 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 21 05:50:54.126482 sshd[1762]: Connection closed by 10.0.0.1 port 53458 Jan 21 05:50:54.134367 sshd-session[1758]: pam_unix(sshd:session): session closed for user core Jan 21 05:50:54.168280 systemd[1]: sshd@1-10.0.0.61:22-10.0.0.1:53458.service: Deactivated successfully. Jan 21 05:50:54.176373 systemd[1]: session-3.scope: Deactivated successfully. Jan 21 05:50:54.181196 systemd-logind[1620]: Session 3 logged out. Waiting for processes to exit. Jan 21 05:50:54.192871 systemd[1]: Started sshd@2-10.0.0.61:22-10.0.0.1:53474.service - OpenSSH per-connection server daemon (10.0.0.1:53474). Jan 21 05:50:54.211277 systemd-logind[1620]: Removed session 3. Jan 21 05:50:54.479873 sshd[1768]: Accepted publickey for core from 10.0.0.1 port 53474 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:50:54.500184 sshd-session[1768]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:50:54.648844 systemd-logind[1620]: New session 4 of user core. Jan 21 05:50:54.674165 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 21 05:50:54.770016 sshd[1773]: Connection closed by 10.0.0.1 port 53474 Jan 21 05:50:54.770928 sshd-session[1768]: pam_unix(sshd:session): session closed for user core Jan 21 05:50:54.800866 systemd[1]: sshd@2-10.0.0.61:22-10.0.0.1:53474.service: Deactivated successfully. Jan 21 05:50:54.811037 systemd[1]: session-4.scope: Deactivated successfully. Jan 21 05:50:54.821283 systemd-logind[1620]: Session 4 logged out. Waiting for processes to exit. Jan 21 05:50:54.827242 systemd-logind[1620]: Removed session 4. Jan 21 05:50:55.747377 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:50:55.754226 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 21 05:50:55.758007 systemd[1]: Startup finished in 15.495s (kernel) + 39.230s (initrd) + 37.728s (userspace) = 1min 32.454s. Jan 21 05:50:55.810406 (kubelet)[1782]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:51:01.277103 kubelet[1782]: E0121 05:51:01.276257 1782 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:51:01.289561 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:51:01.290197 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:51:01.295116 systemd[1]: kubelet.service: Consumed 7.632s CPU time, 271.3M memory peak. Jan 21 05:51:04.850183 systemd[1]: Started sshd@3-10.0.0.61:22-10.0.0.1:52586.service - OpenSSH per-connection server daemon (10.0.0.1:52586). Jan 21 05:51:05.154545 sshd[1793]: Accepted publickey for core from 10.0.0.1 port 52586 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:51:05.160219 sshd-session[1793]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:51:05.211557 systemd-logind[1620]: New session 5 of user core. Jan 21 05:51:05.233192 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 21 05:51:05.349620 sshd[1797]: Connection closed by 10.0.0.1 port 52586 Jan 21 05:51:05.351305 sshd-session[1793]: pam_unix(sshd:session): session closed for user core Jan 21 05:51:05.380110 systemd[1]: sshd@3-10.0.0.61:22-10.0.0.1:52586.service: Deactivated successfully. Jan 21 05:51:05.386223 systemd[1]: session-5.scope: Deactivated successfully. Jan 21 05:51:05.390051 systemd-logind[1620]: Session 5 logged out. Waiting for processes to exit. Jan 21 05:51:05.396060 systemd-logind[1620]: Removed session 5. Jan 21 05:51:05.401546 systemd[1]: Started sshd@4-10.0.0.61:22-10.0.0.1:52594.service - OpenSSH per-connection server daemon (10.0.0.1:52594). Jan 21 05:51:05.606070 sshd[1803]: Accepted publickey for core from 10.0.0.1 port 52594 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:51:05.609192 sshd-session[1803]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:51:05.654095 systemd-logind[1620]: New session 6 of user core. Jan 21 05:51:05.658498 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 21 05:51:05.705113 sshd[1807]: Connection closed by 10.0.0.1 port 52594 Jan 21 05:51:05.703961 sshd-session[1803]: pam_unix(sshd:session): session closed for user core Jan 21 05:51:05.728382 systemd[1]: sshd@4-10.0.0.61:22-10.0.0.1:52594.service: Deactivated successfully. Jan 21 05:51:05.736991 systemd[1]: session-6.scope: Deactivated successfully. Jan 21 05:51:05.758957 systemd-logind[1620]: Session 6 logged out. Waiting for processes to exit. Jan 21 05:51:05.775479 systemd[1]: Started sshd@5-10.0.0.61:22-10.0.0.1:52606.service - OpenSSH per-connection server daemon (10.0.0.1:52606). Jan 21 05:51:05.778918 systemd-logind[1620]: Removed session 6. Jan 21 05:51:06.028145 sshd[1813]: Accepted publickey for core from 10.0.0.1 port 52606 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:51:06.041330 sshd-session[1813]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:51:06.081966 systemd-logind[1620]: New session 7 of user core. Jan 21 05:51:06.107868 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 21 05:51:06.208068 sshd[1817]: Connection closed by 10.0.0.1 port 52606 Jan 21 05:51:06.213287 sshd-session[1813]: pam_unix(sshd:session): session closed for user core Jan 21 05:51:06.260448 systemd[1]: Started sshd@6-10.0.0.61:22-10.0.0.1:52620.service - OpenSSH per-connection server daemon (10.0.0.1:52620). Jan 21 05:51:06.261507 systemd[1]: sshd@5-10.0.0.61:22-10.0.0.1:52606.service: Deactivated successfully. Jan 21 05:51:06.268216 systemd[1]: session-7.scope: Deactivated successfully. Jan 21 05:51:06.272982 systemd-logind[1620]: Session 7 logged out. Waiting for processes to exit. Jan 21 05:51:06.286880 systemd-logind[1620]: Removed session 7. Jan 21 05:51:06.541465 sshd[1820]: Accepted publickey for core from 10.0.0.1 port 52620 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:51:06.561611 sshd-session[1820]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:51:06.604546 systemd-logind[1620]: New session 8 of user core. Jan 21 05:51:06.620488 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 21 05:51:06.784436 sudo[1829]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 21 05:51:06.785306 sudo[1829]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 21 05:51:06.820087 sudo[1829]: pam_unix(sudo:session): session closed for user root Jan 21 05:51:06.826862 sshd[1828]: Connection closed by 10.0.0.1 port 52620 Jan 21 05:51:06.827825 sshd-session[1820]: pam_unix(sshd:session): session closed for user core Jan 21 05:51:06.850219 systemd[1]: sshd@6-10.0.0.61:22-10.0.0.1:52620.service: Deactivated successfully. Jan 21 05:51:06.860904 systemd[1]: session-8.scope: Deactivated successfully. Jan 21 05:51:06.865391 systemd-logind[1620]: Session 8 logged out. Waiting for processes to exit. Jan 21 05:51:06.869846 systemd[1]: Started sshd@7-10.0.0.61:22-10.0.0.1:52632.service - OpenSSH per-connection server daemon (10.0.0.1:52632). Jan 21 05:51:06.877363 systemd-logind[1620]: Removed session 8. Jan 21 05:51:07.385033 sshd[1836]: Accepted publickey for core from 10.0.0.1 port 52632 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:51:07.394431 sshd-session[1836]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:51:07.681238 systemd-logind[1620]: New session 9 of user core. Jan 21 05:51:07.704582 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 21 05:51:07.807411 sudo[1842]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 21 05:51:07.808230 sudo[1842]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 21 05:51:07.829617 sudo[1842]: pam_unix(sudo:session): session closed for user root Jan 21 05:51:07.877061 sudo[1841]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jan 21 05:51:07.878167 sudo[1841]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 21 05:51:07.931363 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 21 05:51:08.256000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 21 05:51:08.260266 augenrules[1866]: No rules Jan 21 05:51:08.263212 kernel: kauditd_printk_skb: 67 callbacks suppressed Jan 21 05:51:08.263309 kernel: audit: type=1305 audit(1768974668.256:216): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 21 05:51:08.268490 systemd[1]: audit-rules.service: Deactivated successfully. Jan 21 05:51:08.269290 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 21 05:51:08.278481 kernel: audit: type=1300 audit(1768974668.256:216): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc5f1814e0 a2=420 a3=0 items=0 ppid=1847 pid=1866 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:08.256000 audit[1866]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc5f1814e0 a2=420 a3=0 items=0 ppid=1847 pid=1866 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:08.279407 sudo[1841]: pam_unix(sudo:session): session closed for user root Jan 21 05:51:08.323477 sshd[1840]: Connection closed by 10.0.0.1 port 52632 Jan 21 05:51:08.324348 kernel: audit: type=1327 audit(1768974668.256:216): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 21 05:51:08.256000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 21 05:51:08.328342 sshd-session[1836]: pam_unix(sshd:session): session closed for user core Jan 21 05:51:08.270000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.430592 kernel: audit: type=1130 audit(1768974668.270:217): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.463919 kernel: audit: type=1131 audit(1768974668.270:218): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.270000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.277000 audit[1841]: USER_END pid=1841 uid=500 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.553309 kernel: audit: type=1106 audit(1768974668.277:219): pid=1841 uid=500 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.553457 kernel: audit: type=1104 audit(1768974668.277:220): pid=1841 uid=500 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.277000 audit[1841]: CRED_DISP pid=1841 uid=500 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.546141 systemd[1]: sshd@7-10.0.0.61:22-10.0.0.1:52632.service: Deactivated successfully. Jan 21 05:51:08.552314 systemd[1]: session-9.scope: Deactivated successfully. Jan 21 05:51:08.565231 systemd-logind[1620]: Session 9 logged out. Waiting for processes to exit. Jan 21 05:51:08.578595 systemd[1]: Started sshd@8-10.0.0.61:22-10.0.0.1:52640.service - OpenSSH per-connection server daemon (10.0.0.1:52640). Jan 21 05:51:08.590139 systemd-logind[1620]: Removed session 9. Jan 21 05:51:08.329000 audit[1836]: USER_END pid=1836 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:51:08.329000 audit[1836]: CRED_DISP pid=1836 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:51:08.661303 kernel: audit: type=1106 audit(1768974668.329:221): pid=1836 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:51:08.661521 kernel: audit: type=1104 audit(1768974668.329:222): pid=1836 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:51:08.661568 kernel: audit: type=1131 audit(1768974668.545:223): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.61:22-10.0.0.1:52632 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.545000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.61:22-10.0.0.1:52632 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.578000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.61:22-10.0.0.1:52640 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:08.916000 audit[1875]: USER_ACCT pid=1875 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:51:08.918817 sshd[1875]: Accepted publickey for core from 10.0.0.1 port 52640 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:51:08.919000 audit[1875]: CRED_ACQ pid=1875 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:51:08.919000 audit[1875]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcfab22430 a2=3 a3=0 items=0 ppid=1 pid=1875 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:08.919000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:51:08.926461 sshd-session[1875]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:51:08.987203 systemd-logind[1620]: New session 10 of user core. Jan 21 05:51:09.029259 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 21 05:51:09.056000 audit[1875]: USER_START pid=1875 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:51:09.063000 audit[1879]: CRED_ACQ pid=1879 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:51:09.121000 audit[1880]: USER_ACCT pid=1880 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:51:09.122000 audit[1880]: CRED_REFR pid=1880 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:51:09.123577 sudo[1880]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 21 05:51:09.124511 sudo[1880]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 21 05:51:09.122000 audit[1880]: USER_START pid=1880 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:51:11.496100 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 21 05:51:11.566121 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:51:14.559773 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 21 05:51:14.594595 (dockerd)[1904]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 21 05:51:14.778826 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:51:14.803757 kernel: kauditd_printk_skb: 11 callbacks suppressed Jan 21 05:51:14.803897 kernel: audit: type=1130 audit(1768974674.778:233): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:14.778000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:14.823742 (kubelet)[1909]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:51:15.063219 kubelet[1909]: E0121 05:51:15.059229 1909 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:51:15.086793 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:51:15.087214 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:51:15.091000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:51:15.091957 systemd[1]: kubelet.service: Consumed 1.960s CPU time, 109.1M memory peak. Jan 21 05:51:15.110895 kernel: audit: type=1131 audit(1768974675.091:234): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:51:15.495302 dockerd[1904]: time="2026-01-21T05:51:15.494520213Z" level=info msg="Starting up" Jan 21 05:51:15.501032 dockerd[1904]: time="2026-01-21T05:51:15.498039869Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Jan 21 05:51:15.564493 dockerd[1904]: time="2026-01-21T05:51:15.564136900Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Jan 21 05:51:15.813837 dockerd[1904]: time="2026-01-21T05:51:15.812304951Z" level=info msg="Loading containers: start." Jan 21 05:51:15.855964 kernel: Initializing XFRM netlink socket Jan 21 05:51:16.100000 audit[1973]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=1973 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.120181 kernel: audit: type=1325 audit(1768974676.100:235): table=nat:2 family=2 entries=2 op=nft_register_chain pid=1973 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.120342 kernel: audit: type=1300 audit(1768974676.100:235): arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffdcaf38f30 a2=0 a3=0 items=0 ppid=1904 pid=1973 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.100000 audit[1973]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffdcaf38f30 a2=0 a3=0 items=0 ppid=1904 pid=1973 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.100000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 21 05:51:16.143767 kernel: audit: type=1327 audit(1768974676.100:235): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 21 05:51:16.143872 kernel: audit: type=1325 audit(1768974676.117:236): table=filter:3 family=2 entries=2 op=nft_register_chain pid=1975 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.117000 audit[1975]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=1975 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.152560 kernel: audit: type=1300 audit(1768974676.117:236): arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffe8c6eb6c0 a2=0 a3=0 items=0 ppid=1904 pid=1975 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.117000 audit[1975]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffe8c6eb6c0 a2=0 a3=0 items=0 ppid=1904 pid=1975 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.117000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 21 05:51:16.174141 kernel: audit: type=1327 audit(1768974676.117:236): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 21 05:51:16.174254 kernel: audit: type=1325 audit(1768974676.129:237): table=filter:4 family=2 entries=1 op=nft_register_chain pid=1977 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.129000 audit[1977]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=1977 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.182893 kernel: audit: type=1300 audit(1768974676.129:237): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcff10d150 a2=0 a3=0 items=0 ppid=1904 pid=1977 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.129000 audit[1977]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcff10d150 a2=0 a3=0 items=0 ppid=1904 pid=1977 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.129000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 21 05:51:16.144000 audit[1979]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=1979 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.144000 audit[1979]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffca37d63f0 a2=0 a3=0 items=0 ppid=1904 pid=1979 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.144000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 21 05:51:16.154000 audit[1981]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_chain pid=1981 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.154000 audit[1981]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffc1640fd90 a2=0 a3=0 items=0 ppid=1904 pid=1981 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.154000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 21 05:51:16.161000 audit[1983]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_chain pid=1983 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.161000 audit[1983]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffdb45f5970 a2=0 a3=0 items=0 ppid=1904 pid=1983 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.161000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 21 05:51:16.170000 audit[1985]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=1985 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.170000 audit[1985]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffc19dc78c0 a2=0 a3=0 items=0 ppid=1904 pid=1985 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.170000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 21 05:51:16.182000 audit[1987]: NETFILTER_CFG table=nat:9 family=2 entries=2 op=nft_register_chain pid=1987 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.182000 audit[1987]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffdf1fe7c50 a2=0 a3=0 items=0 ppid=1904 pid=1987 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.182000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 21 05:51:16.290000 audit[1990]: NETFILTER_CFG table=nat:10 family=2 entries=2 op=nft_register_chain pid=1990 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.290000 audit[1990]: SYSCALL arch=c000003e syscall=46 success=yes exit=472 a0=3 a1=7ffd52b47470 a2=0 a3=0 items=0 ppid=1904 pid=1990 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.290000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 Jan 21 05:51:16.298000 audit[1992]: NETFILTER_CFG table=filter:11 family=2 entries=2 op=nft_register_chain pid=1992 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.298000 audit[1992]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffce5e96530 a2=0 a3=0 items=0 ppid=1904 pid=1992 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.298000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 21 05:51:16.304000 audit[1994]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=1994 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.304000 audit[1994]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffe6aae9670 a2=0 a3=0 items=0 ppid=1904 pid=1994 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.304000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 21 05:51:16.312000 audit[1996]: NETFILTER_CFG table=filter:13 family=2 entries=1 op=nft_register_rule pid=1996 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.312000 audit[1996]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7fff39d5c160 a2=0 a3=0 items=0 ppid=1904 pid=1996 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.312000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 21 05:51:16.320000 audit[1998]: NETFILTER_CFG table=filter:14 family=2 entries=1 op=nft_register_rule pid=1998 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:16.320000 audit[1998]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7ffe94ad6dc0 a2=0 a3=0 items=0 ppid=1904 pid=1998 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.320000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 21 05:51:16.479000 audit[2028]: NETFILTER_CFG table=nat:15 family=10 entries=2 op=nft_register_chain pid=2028 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.479000 audit[2028]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffe92b47360 a2=0 a3=0 items=0 ppid=1904 pid=2028 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.479000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 21 05:51:16.485000 audit[2030]: NETFILTER_CFG table=filter:16 family=10 entries=2 op=nft_register_chain pid=2030 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.485000 audit[2030]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffd2908ddc0 a2=0 a3=0 items=0 ppid=1904 pid=2030 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.485000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 21 05:51:16.491000 audit[2032]: NETFILTER_CFG table=filter:17 family=10 entries=1 op=nft_register_chain pid=2032 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.491000 audit[2032]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffdbdf70ac0 a2=0 a3=0 items=0 ppid=1904 pid=2032 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.491000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 21 05:51:16.499000 audit[2034]: NETFILTER_CFG table=filter:18 family=10 entries=1 op=nft_register_chain pid=2034 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.499000 audit[2034]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffea80a44f0 a2=0 a3=0 items=0 ppid=1904 pid=2034 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.499000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 21 05:51:16.504000 audit[2036]: NETFILTER_CFG table=filter:19 family=10 entries=1 op=nft_register_chain pid=2036 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.504000 audit[2036]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffd1443b290 a2=0 a3=0 items=0 ppid=1904 pid=2036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.504000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 21 05:51:16.517000 audit[2038]: NETFILTER_CFG table=filter:20 family=10 entries=1 op=nft_register_chain pid=2038 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.517000 audit[2038]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffca3b65f80 a2=0 a3=0 items=0 ppid=1904 pid=2038 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.517000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 21 05:51:16.529000 audit[2040]: NETFILTER_CFG table=filter:21 family=10 entries=1 op=nft_register_chain pid=2040 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.529000 audit[2040]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fffc4729f00 a2=0 a3=0 items=0 ppid=1904 pid=2040 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.529000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 21 05:51:16.544000 audit[2042]: NETFILTER_CFG table=nat:22 family=10 entries=2 op=nft_register_chain pid=2042 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.544000 audit[2042]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffd85658d70 a2=0 a3=0 items=0 ppid=1904 pid=2042 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.544000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 21 05:51:16.971000 audit[2044]: NETFILTER_CFG table=nat:23 family=10 entries=2 op=nft_register_chain pid=2044 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.971000 audit[2044]: SYSCALL arch=c000003e syscall=46 success=yes exit=484 a0=3 a1=7ffcf5606630 a2=0 a3=0 items=0 ppid=1904 pid=2044 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.971000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003A3A312F313238 Jan 21 05:51:16.981000 audit[2046]: NETFILTER_CFG table=filter:24 family=10 entries=2 op=nft_register_chain pid=2046 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.981000 audit[2046]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7fff9033ff90 a2=0 a3=0 items=0 ppid=1904 pid=2046 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.981000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 21 05:51:16.991000 audit[2048]: NETFILTER_CFG table=filter:25 family=10 entries=1 op=nft_register_rule pid=2048 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:16.991000 audit[2048]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffc838b7420 a2=0 a3=0 items=0 ppid=1904 pid=2048 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:16.991000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 21 05:51:17.002000 audit[2050]: NETFILTER_CFG table=filter:26 family=10 entries=1 op=nft_register_rule pid=2050 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:17.002000 audit[2050]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7ffeac11f6f0 a2=0 a3=0 items=0 ppid=1904 pid=2050 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.002000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 21 05:51:17.016000 audit[2052]: NETFILTER_CFG table=filter:27 family=10 entries=1 op=nft_register_rule pid=2052 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:17.016000 audit[2052]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7ffe13e32f60 a2=0 a3=0 items=0 ppid=1904 pid=2052 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.016000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 21 05:51:17.037000 audit[2057]: NETFILTER_CFG table=filter:28 family=2 entries=1 op=nft_register_chain pid=2057 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.037000 audit[2057]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffcf6bd54f0 a2=0 a3=0 items=0 ppid=1904 pid=2057 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.037000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 21 05:51:17.067000 audit[2059]: NETFILTER_CFG table=filter:29 family=2 entries=1 op=nft_register_rule pid=2059 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.067000 audit[2059]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffc15954a70 a2=0 a3=0 items=0 ppid=1904 pid=2059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.067000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 21 05:51:17.077000 audit[2061]: NETFILTER_CFG table=filter:30 family=2 entries=1 op=nft_register_rule pid=2061 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.077000 audit[2061]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffe74878060 a2=0 a3=0 items=0 ppid=1904 pid=2061 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.077000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 21 05:51:17.097000 audit[2063]: NETFILTER_CFG table=filter:31 family=10 entries=1 op=nft_register_chain pid=2063 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:17.097000 audit[2063]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffcf7279900 a2=0 a3=0 items=0 ppid=1904 pid=2063 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.097000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 21 05:51:17.110000 audit[2065]: NETFILTER_CFG table=filter:32 family=10 entries=1 op=nft_register_rule pid=2065 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:17.110000 audit[2065]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffe42ae84d0 a2=0 a3=0 items=0 ppid=1904 pid=2065 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.110000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 21 05:51:17.122000 audit[2067]: NETFILTER_CFG table=filter:33 family=10 entries=1 op=nft_register_rule pid=2067 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:51:17.122000 audit[2067]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffccbb995c0 a2=0 a3=0 items=0 ppid=1904 pid=2067 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.122000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 21 05:51:17.238000 audit[2072]: NETFILTER_CFG table=nat:34 family=2 entries=2 op=nft_register_chain pid=2072 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.238000 audit[2072]: SYSCALL arch=c000003e syscall=46 success=yes exit=520 a0=3 a1=7ffe689fbd40 a2=0 a3=0 items=0 ppid=1904 pid=2072 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.238000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 Jan 21 05:51:17.281000 audit[2074]: NETFILTER_CFG table=nat:35 family=2 entries=1 op=nft_register_rule pid=2074 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.281000 audit[2074]: SYSCALL arch=c000003e syscall=46 success=yes exit=288 a0=3 a1=7ffe7d848b30 a2=0 a3=0 items=0 ppid=1904 pid=2074 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.281000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E Jan 21 05:51:17.438000 audit[2082]: NETFILTER_CFG table=filter:36 family=2 entries=1 op=nft_register_rule pid=2082 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.438000 audit[2082]: SYSCALL arch=c000003e syscall=46 success=yes exit=300 a0=3 a1=7ffc390ef7d0 a2=0 a3=0 items=0 ppid=1904 pid=2082 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.438000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D464F5257415244002D6900646F636B657230002D6A00414343455054 Jan 21 05:51:17.530000 audit[2088]: NETFILTER_CFG table=filter:37 family=2 entries=1 op=nft_register_rule pid=2088 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.530000 audit[2088]: SYSCALL arch=c000003e syscall=46 success=yes exit=376 a0=3 a1=7ffc0fdfae80 a2=0 a3=0 items=0 ppid=1904 pid=2088 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.530000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45520000002D6900646F636B657230002D6F00646F636B657230002D6A0044524F50 Jan 21 05:51:17.549000 audit[2090]: NETFILTER_CFG table=filter:38 family=2 entries=1 op=nft_register_rule pid=2090 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.549000 audit[2090]: SYSCALL arch=c000003e syscall=46 success=yes exit=512 a0=3 a1=7ffe75259010 a2=0 a3=0 items=0 ppid=1904 pid=2090 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.549000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D4354002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 Jan 21 05:51:17.560000 audit[2092]: NETFILTER_CFG table=filter:39 family=2 entries=1 op=nft_register_rule pid=2092 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.560000 audit[2092]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffc33d4c730 a2=0 a3=0 items=0 ppid=1904 pid=2092 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.560000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D425249444745002D6F00646F636B657230002D6A00444F434B4552 Jan 21 05:51:17.574000 audit[2094]: NETFILTER_CFG table=filter:40 family=2 entries=1 op=nft_register_rule pid=2094 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.574000 audit[2094]: SYSCALL arch=c000003e syscall=46 success=yes exit=428 a0=3 a1=7ffc73f48e30 a2=0 a3=0 items=0 ppid=1904 pid=2094 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.574000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 21 05:51:17.580000 audit[2096]: NETFILTER_CFG table=filter:41 family=2 entries=1 op=nft_register_rule pid=2096 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:51:17.580000 audit[2096]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffdadbcf670 a2=0 a3=0 items=0 ppid=1904 pid=2096 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:51:17.580000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 Jan 21 05:51:17.583880 systemd-networkd[1528]: docker0: Link UP Jan 21 05:51:17.608459 dockerd[1904]: time="2026-01-21T05:51:17.608304537Z" level=info msg="Loading containers: done." Jan 21 05:51:17.764958 dockerd[1904]: time="2026-01-21T05:51:17.764805021Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 21 05:51:17.765257 dockerd[1904]: time="2026-01-21T05:51:17.764990982Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Jan 21 05:51:17.765257 dockerd[1904]: time="2026-01-21T05:51:17.765203173Z" level=info msg="Initializing buildkit" Jan 21 05:51:18.286779 dockerd[1904]: time="2026-01-21T05:51:18.286176272Z" level=info msg="Completed buildkit initialization" Jan 21 05:51:18.307345 dockerd[1904]: time="2026-01-21T05:51:18.306131940Z" level=info msg="Daemon has completed initialization" Jan 21 05:51:18.307345 dockerd[1904]: time="2026-01-21T05:51:18.306277374Z" level=info msg="API listen on /run/docker.sock" Jan 21 05:51:18.310971 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 21 05:51:18.324000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:24.026967 containerd[1639]: time="2026-01-21T05:51:24.024366210Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.7\"" Jan 21 05:51:25.234463 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 21 05:51:25.260134 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:51:26.347143 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1677489256.mount: Deactivated successfully. Jan 21 05:51:26.906000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:26.902908 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:51:26.943943 kernel: kauditd_printk_skb: 113 callbacks suppressed Jan 21 05:51:26.945061 kernel: audit: type=1130 audit(1768974686.906:276): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:26.997615 (kubelet)[2164]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:51:27.584516 kubelet[2164]: E0121 05:51:27.584038 2164 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:51:27.592488 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:51:27.595023 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:51:27.596126 systemd[1]: kubelet.service: Consumed 1.304s CPU time, 110.8M memory peak. Jan 21 05:51:27.595000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:51:27.613796 kernel: audit: type=1131 audit(1768974687.595:277): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:51:29.324242 update_engine[1626]: I20260121 05:51:29.319190 1626 update_attempter.cc:509] Updating boot flags... Jan 21 05:51:36.197139 containerd[1639]: time="2026-01-21T05:51:36.194043086Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:51:36.197139 containerd[1639]: time="2026-01-21T05:51:36.194842804Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.7: active requests=0, bytes read=30103808" Jan 21 05:51:36.210908 containerd[1639]: time="2026-01-21T05:51:36.208803553Z" level=info msg="ImageCreate event name:\"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:51:36.230471 containerd[1639]: time="2026-01-21T05:51:36.226594421Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:9585226cb85d1dc0f0ef5f7a75f04e4bc91ddd82de249533bd293aa3cf958dab\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:51:36.232754 containerd[1639]: time="2026-01-21T05:51:36.231376095Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.7\" with image id \"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.7\", repo digest \"registry.k8s.io/kube-apiserver@sha256:9585226cb85d1dc0f0ef5f7a75f04e4bc91ddd82de249533bd293aa3cf958dab\", size \"30111311\" in 12.206405679s" Jan 21 05:51:36.235762 containerd[1639]: time="2026-01-21T05:51:36.233205984Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.7\" returns image reference \"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\"" Jan 21 05:51:36.250447 containerd[1639]: time="2026-01-21T05:51:36.249985432Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.7\"" Jan 21 05:51:37.807280 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Jan 21 05:51:37.820834 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:51:39.158000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:39.159020 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:51:39.189952 kernel: audit: type=1130 audit(1768974699.158:278): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:39.200414 (kubelet)[2242]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:51:39.706456 kubelet[2242]: E0121 05:51:39.703857 2242 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:51:39.773000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:51:39.755620 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:51:39.756372 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:51:39.775121 systemd[1]: kubelet.service: Consumed 1.130s CPU time, 110.5M memory peak. Jan 21 05:51:39.807065 kernel: audit: type=1131 audit(1768974699.773:279): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:51:48.173192 containerd[1639]: time="2026-01-21T05:51:48.171157371Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:51:48.182498 containerd[1639]: time="2026-01-21T05:51:48.182384423Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.7: active requests=0, bytes read=26008626" Jan 21 05:51:48.190933 containerd[1639]: time="2026-01-21T05:51:48.190867778Z" level=info msg="ImageCreate event name:\"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:51:48.208189 containerd[1639]: time="2026-01-21T05:51:48.206038487Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f69d77ca0626b5a4b7b432c18de0952941181db7341c80eb89731f46d1d0c230\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:51:48.220169 containerd[1639]: time="2026-01-21T05:51:48.208165617Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.7\" with image id \"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.7\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f69d77ca0626b5a4b7b432c18de0952941181db7341c80eb89731f46d1d0c230\", size \"27673815\" in 11.958120523s" Jan 21 05:51:48.264595 containerd[1639]: time="2026-01-21T05:51:48.232953550Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.7\" returns image reference \"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\"" Jan 21 05:51:48.309913 containerd[1639]: time="2026-01-21T05:51:48.299438093Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.7\"" Jan 21 05:51:50.016143 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Jan 21 05:51:50.046437 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:51:51.228341 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:51:51.232000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:51.269608 kernel: audit: type=1130 audit(1768974711.232:280): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:51:51.289515 (kubelet)[2264]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:51:51.874587 kubelet[2264]: E0121 05:51:51.874045 2264 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:51:51.884839 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:51:51.886232 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:51:51.891052 systemd[1]: kubelet.service: Consumed 842ms CPU time, 110.6M memory peak. Jan 21 05:51:51.890000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:51:51.916708 kernel: audit: type=1131 audit(1768974711.890:281): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:51:57.278859 containerd[1639]: time="2026-01-21T05:51:57.277499330Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:51:57.286859 containerd[1639]: time="2026-01-21T05:51:57.281868478Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.7: active requests=0, bytes read=20149965" Jan 21 05:51:57.286859 containerd[1639]: time="2026-01-21T05:51:57.286250705Z" level=info msg="ImageCreate event name:\"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:51:57.304149 containerd[1639]: time="2026-01-21T05:51:57.302528403Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:21bda321d8b4d48eb059fbc1593203d55d8b3bc7acd0584e04e55504796d78d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:51:57.309828 containerd[1639]: time="2026-01-21T05:51:57.307493209Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.7\" with image id \"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.7\", repo digest \"registry.k8s.io/kube-scheduler@sha256:21bda321d8b4d48eb059fbc1593203d55d8b3bc7acd0584e04e55504796d78d0\", size \"21815154\" in 8.993156638s" Jan 21 05:51:57.309828 containerd[1639]: time="2026-01-21T05:51:57.308848952Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.7\" returns image reference \"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\"" Jan 21 05:51:57.318170 containerd[1639]: time="2026-01-21T05:51:57.317470760Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.7\"" Jan 21 05:52:02.046444 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Jan 21 05:52:02.093315 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:52:03.279811 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:52:03.277000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:03.371370 kernel: audit: type=1130 audit(1768974723.277:282): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:03.387771 (kubelet)[2285]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:52:03.677301 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3203583118.mount: Deactivated successfully. Jan 21 05:52:04.078828 kubelet[2285]: E0121 05:52:04.078262 2285 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:52:04.093255 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:52:04.094939 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:52:04.096956 systemd[1]: kubelet.service: Consumed 905ms CPU time, 108.3M memory peak. Jan 21 05:52:04.096000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:04.138796 kernel: audit: type=1131 audit(1768974724.096:283): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:11.376463 containerd[1639]: time="2026-01-21T05:52:11.375570990Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:52:11.386152 containerd[1639]: time="2026-01-21T05:52:11.385801844Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.7: active requests=0, bytes read=31927114" Jan 21 05:52:11.401522 containerd[1639]: time="2026-01-21T05:52:11.401201689Z" level=info msg="ImageCreate event name:\"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:52:11.423227 containerd[1639]: time="2026-01-21T05:52:11.418810587Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ec25702b19026e9c0d339bc1c3bd231435a59f28b5fccb21e1b1078a357380f5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:52:11.423227 containerd[1639]: time="2026-01-21T05:52:11.420030377Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.7\" with image id \"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\", repo tag \"registry.k8s.io/kube-proxy:v1.33.7\", repo digest \"registry.k8s.io/kube-proxy@sha256:ec25702b19026e9c0d339bc1c3bd231435a59f28b5fccb21e1b1078a357380f5\", size \"31929115\" in 14.102400941s" Jan 21 05:52:11.423227 containerd[1639]: time="2026-01-21T05:52:11.420072677Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.7\" returns image reference \"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\"" Jan 21 05:52:11.436088 containerd[1639]: time="2026-01-21T05:52:11.435218324Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Jan 21 05:52:14.162301 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount7920219.mount: Deactivated successfully. Jan 21 05:52:14.165535 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Jan 21 05:52:14.180196 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:52:15.699352 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:52:15.700000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:15.735879 kernel: audit: type=1130 audit(1768974735.700:284): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:15.755384 (kubelet)[2317]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:52:16.351763 kubelet[2317]: E0121 05:52:16.346062 2317 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:52:16.360240 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:52:16.360735 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:52:16.367479 systemd[1]: kubelet.service: Consumed 1.061s CPU time, 110.8M memory peak. Jan 21 05:52:16.366000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:16.427135 kernel: audit: type=1131 audit(1768974736.366:285): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:23.452209 containerd[1639]: time="2026-01-21T05:52:23.447036323Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:52:23.521148 containerd[1639]: time="2026-01-21T05:52:23.469121469Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20934390" Jan 21 05:52:23.571483 containerd[1639]: time="2026-01-21T05:52:23.569422161Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:52:23.714744 containerd[1639]: time="2026-01-21T05:52:23.704600716Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:52:23.714744 containerd[1639]: time="2026-01-21T05:52:23.706564055Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 12.271196157s" Jan 21 05:52:23.714744 containerd[1639]: time="2026-01-21T05:52:23.706605693Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Jan 21 05:52:23.722748 containerd[1639]: time="2026-01-21T05:52:23.719961263Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jan 21 05:52:24.890069 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3798783949.mount: Deactivated successfully. Jan 21 05:52:24.921978 containerd[1639]: time="2026-01-21T05:52:24.921814749Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 21 05:52:24.925593 containerd[1639]: time="2026-01-21T05:52:24.924849486Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Jan 21 05:52:24.928279 containerd[1639]: time="2026-01-21T05:52:24.927329741Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 21 05:52:24.931620 containerd[1639]: time="2026-01-21T05:52:24.931329983Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 21 05:52:24.935316 containerd[1639]: time="2026-01-21T05:52:24.932567048Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 1.212479068s" Jan 21 05:52:24.935316 containerd[1639]: time="2026-01-21T05:52:24.934113034Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Jan 21 05:52:24.941993 containerd[1639]: time="2026-01-21T05:52:24.940940986Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Jan 21 05:52:26.489526 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Jan 21 05:52:26.521151 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:52:27.491131 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3212823732.mount: Deactivated successfully. Jan 21 05:52:27.599870 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:52:27.630360 kernel: audit: type=1130 audit(1768974747.600:286): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:27.600000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:27.842097 (kubelet)[2381]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:52:29.600908 kubelet[2381]: E0121 05:52:29.600130 2381 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:52:29.620822 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:52:29.632838 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:52:29.638000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:29.688621 systemd[1]: kubelet.service: Consumed 1.321s CPU time, 108.1M memory peak. Jan 21 05:52:29.701800 kernel: audit: type=1131 audit(1768974749.638:287): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:39.772347 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Jan 21 05:52:39.790878 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:52:40.810160 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:52:40.809000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:40.837778 kernel: audit: type=1130 audit(1768974760.809:288): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:40.856148 (kubelet)[2445]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:52:41.659805 kubelet[2445]: E0121 05:52:41.657891 2445 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:52:41.671161 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:52:41.671629 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:52:41.673282 systemd[1]: kubelet.service: Consumed 903ms CPU time, 112.6M memory peak. Jan 21 05:52:41.672000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:41.699055 kernel: audit: type=1131 audit(1768974761.672:289): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:46.117310 containerd[1639]: time="2026-01-21T05:52:46.116077731Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:52:46.137301 containerd[1639]: time="2026-01-21T05:52:46.136894488Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58916088" Jan 21 05:52:46.144771 containerd[1639]: time="2026-01-21T05:52:46.144411747Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:52:46.151474 containerd[1639]: time="2026-01-21T05:52:46.150954924Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:52:46.157457 containerd[1639]: time="2026-01-21T05:52:46.153526717Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 21.21254277s" Jan 21 05:52:46.157457 containerd[1639]: time="2026-01-21T05:52:46.153585599Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Jan 21 05:52:51.820452 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Jan 21 05:52:51.844353 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:52:53.270000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:53.272095 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:52:53.292911 kernel: audit: type=1130 audit(1768974773.270:290): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:53.325289 (kubelet)[2492]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 21 05:52:53.625500 kubelet[2492]: E0121 05:52:53.624983 2492 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 21 05:52:53.643000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:53.643044 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 21 05:52:53.643307 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 21 05:52:53.644465 systemd[1]: kubelet.service: Consumed 772ms CPU time, 110.4M memory peak. Jan 21 05:52:53.697334 kernel: audit: type=1131 audit(1768974773.643:291): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:54.400000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:54.401949 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:52:54.402283 systemd[1]: kubelet.service: Consumed 772ms CPU time, 110.4M memory peak. Jan 21 05:52:54.410364 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:52:54.429850 kernel: audit: type=1130 audit(1768974774.400:292): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:54.400000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:54.456122 kernel: audit: type=1131 audit(1768974774.400:293): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:54.508233 systemd[1]: Reload requested from client PID 2508 ('systemctl') (unit session-10.scope)... Jan 21 05:52:54.508307 systemd[1]: Reloading... Jan 21 05:52:54.734890 zram_generator::config[2550]: No configuration found. Jan 21 05:52:55.422204 systemd[1]: Reloading finished in 912 ms. Jan 21 05:52:55.509000 audit: BPF prog-id=61 op=LOAD Jan 21 05:52:55.531428 kernel: audit: type=1334 audit(1768974775.509:294): prog-id=61 op=LOAD Jan 21 05:52:55.531545 kernel: audit: type=1334 audit(1768974775.509:295): prog-id=48 op=UNLOAD Jan 21 05:52:55.509000 audit: BPF prog-id=48 op=UNLOAD Jan 21 05:52:55.509000 audit: BPF prog-id=62 op=LOAD Jan 21 05:52:55.509000 audit: BPF prog-id=63 op=LOAD Jan 21 05:52:55.553096 kernel: audit: type=1334 audit(1768974775.509:296): prog-id=62 op=LOAD Jan 21 05:52:55.553206 kernel: audit: type=1334 audit(1768974775.509:297): prog-id=63 op=LOAD Jan 21 05:52:55.553316 kernel: audit: type=1334 audit(1768974775.509:298): prog-id=49 op=UNLOAD Jan 21 05:52:55.509000 audit: BPF prog-id=49 op=UNLOAD Jan 21 05:52:55.564108 kernel: audit: type=1334 audit(1768974775.509:299): prog-id=50 op=UNLOAD Jan 21 05:52:55.509000 audit: BPF prog-id=50 op=UNLOAD Jan 21 05:52:55.513000 audit: BPF prog-id=64 op=LOAD Jan 21 05:52:55.513000 audit: BPF prog-id=47 op=UNLOAD Jan 21 05:52:55.517000 audit: BPF prog-id=65 op=LOAD Jan 21 05:52:55.517000 audit: BPF prog-id=51 op=UNLOAD Jan 21 05:52:55.517000 audit: BPF prog-id=66 op=LOAD Jan 21 05:52:55.517000 audit: BPF prog-id=67 op=LOAD Jan 21 05:52:55.517000 audit: BPF prog-id=52 op=UNLOAD Jan 21 05:52:55.517000 audit: BPF prog-id=53 op=UNLOAD Jan 21 05:52:55.520000 audit: BPF prog-id=68 op=LOAD Jan 21 05:52:55.521000 audit: BPF prog-id=69 op=LOAD Jan 21 05:52:55.523000 audit: BPF prog-id=54 op=UNLOAD Jan 21 05:52:55.523000 audit: BPF prog-id=55 op=UNLOAD Jan 21 05:52:55.525000 audit: BPF prog-id=70 op=LOAD Jan 21 05:52:55.525000 audit: BPF prog-id=41 op=UNLOAD Jan 21 05:52:55.525000 audit: BPF prog-id=71 op=LOAD Jan 21 05:52:55.526000 audit: BPF prog-id=72 op=LOAD Jan 21 05:52:55.527000 audit: BPF prog-id=42 op=UNLOAD Jan 21 05:52:55.527000 audit: BPF prog-id=43 op=UNLOAD Jan 21 05:52:55.529000 audit: BPF prog-id=73 op=LOAD Jan 21 05:52:55.529000 audit: BPF prog-id=56 op=UNLOAD Jan 21 05:52:55.537000 audit: BPF prog-id=74 op=LOAD Jan 21 05:52:55.573000 audit: BPF prog-id=58 op=UNLOAD Jan 21 05:52:55.573000 audit: BPF prog-id=75 op=LOAD Jan 21 05:52:55.573000 audit: BPF prog-id=76 op=LOAD Jan 21 05:52:55.573000 audit: BPF prog-id=59 op=UNLOAD Jan 21 05:52:55.573000 audit: BPF prog-id=60 op=UNLOAD Jan 21 05:52:55.573000 audit: BPF prog-id=77 op=LOAD Jan 21 05:52:55.573000 audit: BPF prog-id=44 op=UNLOAD Jan 21 05:52:55.573000 audit: BPF prog-id=78 op=LOAD Jan 21 05:52:55.573000 audit: BPF prog-id=79 op=LOAD Jan 21 05:52:55.573000 audit: BPF prog-id=45 op=UNLOAD Jan 21 05:52:55.573000 audit: BPF prog-id=46 op=UNLOAD Jan 21 05:52:55.581000 audit: BPF prog-id=80 op=LOAD Jan 21 05:52:55.581000 audit: BPF prog-id=57 op=UNLOAD Jan 21 05:52:55.673225 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 21 05:52:55.673341 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 21 05:52:55.675131 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:52:55.675390 systemd[1]: kubelet.service: Consumed 234ms CPU time, 98.7M memory peak. Jan 21 05:52:55.673000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 21 05:52:55.682172 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:52:56.139232 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:52:56.138000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:52:56.202930 (kubelet)[2602]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 21 05:52:57.182248 kubelet[2602]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 05:52:57.182248 kubelet[2602]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 21 05:52:57.182248 kubelet[2602]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 05:52:57.182248 kubelet[2602]: I0121 05:52:57.183455 2602 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 21 05:52:58.643475 kubelet[2602]: I0121 05:52:58.643011 2602 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jan 21 05:52:58.643475 kubelet[2602]: I0121 05:52:58.643475 2602 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 21 05:52:58.646226 kubelet[2602]: I0121 05:52:58.646128 2602 server.go:956] "Client rotation is on, will bootstrap in background" Jan 21 05:52:58.802329 kubelet[2602]: E0121 05:52:58.800455 2602 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.61:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 21 05:52:58.802329 kubelet[2602]: I0121 05:52:58.800610 2602 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 21 05:52:58.855729 kubelet[2602]: I0121 05:52:58.855359 2602 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 21 05:52:58.886609 kubelet[2602]: I0121 05:52:58.884580 2602 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 21 05:52:58.886609 kubelet[2602]: I0121 05:52:58.886026 2602 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 21 05:52:58.886609 kubelet[2602]: I0121 05:52:58.886084 2602 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 21 05:52:58.888236 kubelet[2602]: I0121 05:52:58.886779 2602 topology_manager.go:138] "Creating topology manager with none policy" Jan 21 05:52:58.888236 kubelet[2602]: I0121 05:52:58.886841 2602 container_manager_linux.go:303] "Creating device plugin manager" Jan 21 05:52:58.888236 kubelet[2602]: I0121 05:52:58.887442 2602 state_mem.go:36] "Initialized new in-memory state store" Jan 21 05:52:58.908924 kubelet[2602]: I0121 05:52:58.908241 2602 kubelet.go:480] "Attempting to sync node with API server" Jan 21 05:52:58.910376 kubelet[2602]: I0121 05:52:58.909799 2602 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 21 05:52:58.910376 kubelet[2602]: I0121 05:52:58.909856 2602 kubelet.go:386] "Adding apiserver pod source" Jan 21 05:52:58.910376 kubelet[2602]: I0121 05:52:58.910045 2602 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 21 05:52:58.922784 kubelet[2602]: E0121 05:52:58.922317 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.61:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 21 05:52:58.938316 kubelet[2602]: I0121 05:52:58.937105 2602 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 21 05:52:58.942253 kubelet[2602]: I0121 05:52:58.940327 2602 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 21 05:52:58.942253 kubelet[2602]: E0121 05:52:58.941557 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.61:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 21 05:52:58.953556 kubelet[2602]: W0121 05:52:58.952453 2602 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 21 05:52:58.977145 kubelet[2602]: I0121 05:52:58.976818 2602 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 21 05:52:58.977145 kubelet[2602]: I0121 05:52:58.977059 2602 server.go:1289] "Started kubelet" Jan 21 05:52:58.978820 kubelet[2602]: I0121 05:52:58.978032 2602 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 21 05:52:58.979203 kubelet[2602]: I0121 05:52:58.979069 2602 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 21 05:52:58.985784 kubelet[2602]: I0121 05:52:58.980439 2602 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 21 05:52:58.985784 kubelet[2602]: I0121 05:52:58.982341 2602 server.go:317] "Adding debug handlers to kubelet server" Jan 21 05:52:58.995550 kubelet[2602]: I0121 05:52:58.994136 2602 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 21 05:52:58.995550 kubelet[2602]: E0121 05:52:58.988112 2602 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.61:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.61:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.188ca930658b252a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-21 05:52:58.97690449 +0000 UTC m=+2.710366281,LastTimestamp:2026-01-21 05:52:58.97690449 +0000 UTC m=+2.710366281,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 21 05:52:58.998754 kubelet[2602]: I0121 05:52:58.996471 2602 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 21 05:52:59.007187 kubelet[2602]: I0121 05:52:59.007086 2602 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 21 05:52:59.013813 kubelet[2602]: E0121 05:52:59.013456 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:52:59.102198 kubelet[2602]: I0121 05:52:59.099812 2602 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jan 21 05:52:59.102198 kubelet[2602]: I0121 05:52:59.101169 2602 reconciler.go:26] "Reconciler: start to sync state" Jan 21 05:52:59.102198 kubelet[2602]: E0121 05:52:59.018614 2602 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.61:6443: connect: connection refused" interval="200ms" Jan 21 05:52:59.111800 kubelet[2602]: E0121 05:52:59.103172 2602 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 21 05:52:59.111800 kubelet[2602]: I0121 05:52:59.103601 2602 factory.go:223] Registration of the systemd container factory successfully Jan 21 05:52:59.111800 kubelet[2602]: I0121 05:52:59.103923 2602 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 21 05:52:59.111800 kubelet[2602]: E0121 05:52:59.104061 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.61:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 21 05:52:59.116803 kubelet[2602]: I0121 05:52:59.115144 2602 factory.go:223] Registration of the containerd container factory successfully Jan 21 05:52:59.128000 audit[2621]: NETFILTER_CFG table=mangle:42 family=2 entries=2 op=nft_register_chain pid=2621 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.140493 kernel: kauditd_printk_skb: 36 callbacks suppressed Jan 21 05:52:59.140952 kernel: audit: type=1325 audit(1768974779.128:336): table=mangle:42 family=2 entries=2 op=nft_register_chain pid=2621 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.156024 kubelet[2602]: E0121 05:52:59.155806 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:52:59.167483 kernel: audit: type=1300 audit(1768974779.128:336): arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7fffb137a020 a2=0 a3=0 items=0 ppid=2602 pid=2621 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.128000 audit[2621]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7fffb137a020 a2=0 a3=0 items=0 ppid=2602 pid=2621 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.128000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 21 05:52:59.241570 kernel: audit: type=1327 audit(1768974779.128:336): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 21 05:52:59.241833 kernel: audit: type=1325 audit(1768974779.200:337): table=filter:43 family=2 entries=1 op=nft_register_chain pid=2622 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.200000 audit[2622]: NETFILTER_CFG table=filter:43 family=2 entries=1 op=nft_register_chain pid=2622 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.259760 kubelet[2602]: E0121 05:52:59.258768 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:52:59.264504 kernel: audit: type=1300 audit(1768974779.200:337): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffce39efbe0 a2=0 a3=0 items=0 ppid=2602 pid=2622 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.200000 audit[2622]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffce39efbe0 a2=0 a3=0 items=0 ppid=2602 pid=2622 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.278338 kubelet[2602]: I0121 05:52:59.278118 2602 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 21 05:52:59.278338 kubelet[2602]: I0121 05:52:59.278201 2602 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 21 05:52:59.278338 kubelet[2602]: I0121 05:52:59.278226 2602 state_mem.go:36] "Initialized new in-memory state store" Jan 21 05:52:59.306906 kubelet[2602]: E0121 05:52:59.305868 2602 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.61:6443: connect: connection refused" interval="400ms" Jan 21 05:52:59.309571 kubelet[2602]: I0121 05:52:59.307224 2602 policy_none.go:49] "None policy: Start" Jan 21 05:52:59.309571 kubelet[2602]: I0121 05:52:59.308324 2602 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 21 05:52:59.309571 kubelet[2602]: I0121 05:52:59.308420 2602 state_mem.go:35] "Initializing new in-memory state store" Jan 21 05:52:59.200000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 21 05:52:59.352159 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 21 05:52:59.380597 kernel: audit: type=1327 audit(1768974779.200:337): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 21 05:52:59.380867 kernel: audit: type=1325 audit(1768974779.271:338): table=filter:44 family=2 entries=2 op=nft_register_chain pid=2628 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.271000 audit[2628]: NETFILTER_CFG table=filter:44 family=2 entries=2 op=nft_register_chain pid=2628 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.380976 kubelet[2602]: E0121 05:52:59.359968 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:52:59.271000 audit[2628]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffe378bd550 a2=0 a3=0 items=0 ppid=2602 pid=2628 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.271000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 21 05:52:59.454949 kernel: audit: type=1300 audit(1768974779.271:338): arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffe378bd550 a2=0 a3=0 items=0 ppid=2602 pid=2628 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.463816 kernel: audit: type=1327 audit(1768974779.271:338): proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 21 05:52:59.463888 kernel: audit: type=1325 audit(1768974779.297:339): table=filter:45 family=2 entries=2 op=nft_register_chain pid=2631 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.297000 audit[2631]: NETFILTER_CFG table=filter:45 family=2 entries=2 op=nft_register_chain pid=2631 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.464820 kubelet[2602]: E0121 05:52:59.461487 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:52:59.297000 audit[2631]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7fff0cb36720 a2=0 a3=0 items=0 ppid=2602 pid=2631 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.297000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 21 05:52:59.482000 audit[2635]: NETFILTER_CFG table=filter:46 family=2 entries=1 op=nft_register_rule pid=2635 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.482000 audit[2635]: SYSCALL arch=c000003e syscall=46 success=yes exit=924 a0=3 a1=7ffe5b378a80 a2=0 a3=0 items=0 ppid=2602 pid=2635 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.482000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F38 Jan 21 05:52:59.486319 kubelet[2602]: I0121 05:52:59.486139 2602 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jan 21 05:52:59.490000 audit[2636]: NETFILTER_CFG table=mangle:47 family=10 entries=2 op=nft_register_chain pid=2636 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:52:59.490000 audit[2636]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffdb7dae630 a2=0 a3=0 items=0 ppid=2602 pid=2636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.490000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 21 05:52:59.491927 kubelet[2602]: I0121 05:52:59.491584 2602 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jan 21 05:52:59.494830 kubelet[2602]: I0121 05:52:59.492204 2602 status_manager.go:230] "Starting to sync pod status with apiserver" Jan 21 05:52:59.494830 kubelet[2602]: I0121 05:52:59.492427 2602 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 21 05:52:59.494830 kubelet[2602]: I0121 05:52:59.492493 2602 kubelet.go:2436] "Starting kubelet main sync loop" Jan 21 05:52:59.494830 kubelet[2602]: E0121 05:52:59.493142 2602 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 21 05:52:59.501000 audit[2638]: NETFILTER_CFG table=mangle:48 family=10 entries=1 op=nft_register_chain pid=2638 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:52:59.501000 audit[2638]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe348ec0f0 a2=0 a3=0 items=0 ppid=2602 pid=2638 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.501000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 21 05:52:59.503040 kubelet[2602]: E0121 05:52:59.502239 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.61:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 21 05:52:59.502931 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 21 05:52:59.517000 audit[2637]: NETFILTER_CFG table=mangle:49 family=2 entries=1 op=nft_register_chain pid=2637 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.517000 audit[2637]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe7a66f610 a2=0 a3=0 items=0 ppid=2602 pid=2637 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.517000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 21 05:52:59.520000 audit[2639]: NETFILTER_CFG table=nat:50 family=10 entries=1 op=nft_register_chain pid=2639 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:52:59.520000 audit[2639]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcd83a66e0 a2=0 a3=0 items=0 ppid=2602 pid=2639 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.520000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 21 05:52:59.526000 audit[2641]: NETFILTER_CFG table=filter:51 family=10 entries=1 op=nft_register_chain pid=2641 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:52:59.526000 audit[2641]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe36b9d2a0 a2=0 a3=0 items=0 ppid=2602 pid=2641 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.526000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 21 05:52:59.536115 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 21 05:52:59.537000 audit[2640]: NETFILTER_CFG table=nat:52 family=2 entries=1 op=nft_register_chain pid=2640 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.537000 audit[2640]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd87223ee0 a2=0 a3=0 items=0 ppid=2602 pid=2640 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.537000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 21 05:52:59.542000 audit[2642]: NETFILTER_CFG table=filter:53 family=2 entries=1 op=nft_register_chain pid=2642 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:52:59.542000 audit[2642]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffcad42e640 a2=0 a3=0 items=0 ppid=2602 pid=2642 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:52:59.542000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 21 05:52:59.563523 kubelet[2602]: E0121 05:52:59.563462 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:52:59.566859 kubelet[2602]: E0121 05:52:59.566582 2602 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 21 05:52:59.568799 kubelet[2602]: I0121 05:52:59.567212 2602 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 21 05:52:59.571088 kubelet[2602]: I0121 05:52:59.567286 2602 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 21 05:52:59.573278 kubelet[2602]: I0121 05:52:59.572068 2602 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 21 05:52:59.576445 kubelet[2602]: E0121 05:52:59.574896 2602 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 21 05:52:59.576445 kubelet[2602]: E0121 05:52:59.575246 2602 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 21 05:52:59.610604 kubelet[2602]: I0121 05:52:59.610201 2602 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/52e4d30ab66d3d1182c212fc42b84441-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"52e4d30ab66d3d1182c212fc42b84441\") " pod="kube-system/kube-apiserver-localhost" Jan 21 05:52:59.610604 kubelet[2602]: I0121 05:52:59.610314 2602 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/52e4d30ab66d3d1182c212fc42b84441-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"52e4d30ab66d3d1182c212fc42b84441\") " pod="kube-system/kube-apiserver-localhost" Jan 21 05:52:59.610604 kubelet[2602]: I0121 05:52:59.610427 2602 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/52e4d30ab66d3d1182c212fc42b84441-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"52e4d30ab66d3d1182c212fc42b84441\") " pod="kube-system/kube-apiserver-localhost" Jan 21 05:52:59.687376 systemd[1]: Created slice kubepods-burstable-pod52e4d30ab66d3d1182c212fc42b84441.slice - libcontainer container kubepods-burstable-pod52e4d30ab66d3d1182c212fc42b84441.slice. Jan 21 05:52:59.695075 kubelet[2602]: I0121 05:52:59.694962 2602 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 21 05:52:59.695983 kubelet[2602]: E0121 05:52:59.695834 2602 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.61:6443/api/v1/nodes\": dial tcp 10.0.0.61:6443: connect: connection refused" node="localhost" Jan 21 05:52:59.707606 kubelet[2602]: E0121 05:52:59.707441 2602 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.61:6443: connect: connection refused" interval="800ms" Jan 21 05:52:59.711615 kubelet[2602]: I0121 05:52:59.711456 2602 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6e6cfcfb327385445a9bb0d2bc2fd5d4-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6e6cfcfb327385445a9bb0d2bc2fd5d4\") " pod="kube-system/kube-scheduler-localhost" Jan 21 05:52:59.711615 kubelet[2602]: I0121 05:52:59.711504 2602 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:52:59.711615 kubelet[2602]: I0121 05:52:59.711563 2602 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:52:59.711615 kubelet[2602]: I0121 05:52:59.711587 2602 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:52:59.711964 kubelet[2602]: I0121 05:52:59.711612 2602 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:52:59.711964 kubelet[2602]: I0121 05:52:59.711760 2602 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:52:59.713455 kubelet[2602]: E0121 05:52:59.713249 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:52:59.719625 kubelet[2602]: E0121 05:52:59.716055 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:52:59.721153 containerd[1639]: time="2026-01-21T05:52:59.720456035Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:52e4d30ab66d3d1182c212fc42b84441,Namespace:kube-system,Attempt:0,}" Jan 21 05:52:59.723550 systemd[1]: Created slice kubepods-burstable-pod66e26b992bcd7ea6fb75e339cf7a3f7d.slice - libcontainer container kubepods-burstable-pod66e26b992bcd7ea6fb75e339cf7a3f7d.slice. Jan 21 05:52:59.747084 kubelet[2602]: E0121 05:52:59.746805 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:52:59.760438 systemd[1]: Created slice kubepods-burstable-pod6e6cfcfb327385445a9bb0d2bc2fd5d4.slice - libcontainer container kubepods-burstable-pod6e6cfcfb327385445a9bb0d2bc2fd5d4.slice. Jan 21 05:52:59.774002 kubelet[2602]: E0121 05:52:59.772105 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:00.016825 kubelet[2602]: I0121 05:53:00.007993 2602 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 21 05:53:00.016825 kubelet[2602]: E0121 05:53:00.014862 2602 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.61:6443/api/v1/nodes\": dial tcp 10.0.0.61:6443: connect: connection refused" node="localhost" Jan 21 05:53:00.072702 kubelet[2602]: E0121 05:53:00.072399 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:00.075617 kubelet[2602]: E0121 05:53:00.075155 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:00.158412 containerd[1639]: time="2026-01-21T05:53:00.147364718Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6e6cfcfb327385445a9bb0d2bc2fd5d4,Namespace:kube-system,Attempt:0,}" Jan 21 05:53:00.170747 containerd[1639]: time="2026-01-21T05:53:00.161233817Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:66e26b992bcd7ea6fb75e339cf7a3f7d,Namespace:kube-system,Attempt:0,}" Jan 21 05:53:00.170747 containerd[1639]: time="2026-01-21T05:53:00.167944905Z" level=info msg="connecting to shim 961c312c87d378139c025716e98fcb3edc7f8ef6c69f561cd7794866bc1be685" address="unix:///run/containerd/s/3a27688d86be83c0433a638d1f3f85749f0c13ee32812c8398e84b0e57b62129" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:53:00.270237 kubelet[2602]: E0121 05:53:00.268363 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.61:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 21 05:53:00.377558 kubelet[2602]: E0121 05:53:00.376850 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.61:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 21 05:53:00.413395 systemd[1]: Started cri-containerd-961c312c87d378139c025716e98fcb3edc7f8ef6c69f561cd7794866bc1be685.scope - libcontainer container 961c312c87d378139c025716e98fcb3edc7f8ef6c69f561cd7794866bc1be685. Jan 21 05:53:00.421602 kubelet[2602]: I0121 05:53:00.420099 2602 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 21 05:53:00.421602 kubelet[2602]: E0121 05:53:00.420764 2602 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.61:6443/api/v1/nodes\": dial tcp 10.0.0.61:6443: connect: connection refused" node="localhost" Jan 21 05:53:00.427789 kubelet[2602]: E0121 05:53:00.427469 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.61:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 21 05:53:00.509864 kubelet[2602]: E0121 05:53:00.509203 2602 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.61:6443: connect: connection refused" interval="1.6s" Jan 21 05:53:00.530000 audit: BPF prog-id=81 op=LOAD Jan 21 05:53:00.532567 containerd[1639]: time="2026-01-21T05:53:00.532520994Z" level=info msg="connecting to shim c05d41a31e0516324b4efaba818e5e7db89659c320c2c67e4323abde449ed506" address="unix:///run/containerd/s/b89edb0b4723cad04b13f0bd2ba5cf81d56f7549127e0bb9b67c48f2bc6de2d3" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:53:00.553000 audit: BPF prog-id=82 op=LOAD Jan 21 05:53:00.553000 audit[2662]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0238 a2=98 a3=0 items=0 ppid=2651 pid=2662 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:00.553000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936316333313263383764333738313339633032353731366539386663 Jan 21 05:53:00.555000 audit: BPF prog-id=82 op=UNLOAD Jan 21 05:53:00.555000 audit[2662]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2651 pid=2662 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:00.555000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936316333313263383764333738313339633032353731366539386663 Jan 21 05:53:00.557000 audit: BPF prog-id=83 op=LOAD Jan 21 05:53:00.557000 audit[2662]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=2651 pid=2662 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:00.557000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936316333313263383764333738313339633032353731366539386663 Jan 21 05:53:00.557000 audit: BPF prog-id=84 op=LOAD Jan 21 05:53:00.557000 audit[2662]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=2651 pid=2662 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:00.557000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936316333313263383764333738313339633032353731366539386663 Jan 21 05:53:00.557000 audit: BPF prog-id=84 op=UNLOAD Jan 21 05:53:00.557000 audit[2662]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2651 pid=2662 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:00.557000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936316333313263383764333738313339633032353731366539386663 Jan 21 05:53:00.557000 audit: BPF prog-id=83 op=UNLOAD Jan 21 05:53:00.557000 audit[2662]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2651 pid=2662 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:00.557000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936316333313263383764333738313339633032353731366539386663 Jan 21 05:53:00.557000 audit: BPF prog-id=85 op=LOAD Jan 21 05:53:00.557000 audit[2662]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=2651 pid=2662 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:00.557000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936316333313263383764333738313339633032353731366539386663 Jan 21 05:53:00.572133 containerd[1639]: time="2026-01-21T05:53:00.571341756Z" level=info msg="connecting to shim c281aa1f00ad182be9b55fb6ba1e0bacf920c343122844348606c6c2ab448bb9" address="unix:///run/containerd/s/4d9e1d05240c8ee62e2c6248539594665496e95f27db63afb582651c4add99c4" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:53:00.595158 kubelet[2602]: E0121 05:53:00.595090 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.61:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 21 05:53:00.994178 systemd[1]: Started cri-containerd-c281aa1f00ad182be9b55fb6ba1e0bacf920c343122844348606c6c2ab448bb9.scope - libcontainer container c281aa1f00ad182be9b55fb6ba1e0bacf920c343122844348606c6c2ab448bb9. Jan 21 05:53:01.022267 kubelet[2602]: E0121 05:53:01.022010 2602 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.61:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 21 05:53:01.067572 containerd[1639]: time="2026-01-21T05:53:01.067442675Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:52e4d30ab66d3d1182c212fc42b84441,Namespace:kube-system,Attempt:0,} returns sandbox id \"961c312c87d378139c025716e98fcb3edc7f8ef6c69f561cd7794866bc1be685\"" Jan 21 05:53:01.070317 kubelet[2602]: E0121 05:53:01.070106 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:01.070794 systemd[1]: Started cri-containerd-c05d41a31e0516324b4efaba818e5e7db89659c320c2c67e4323abde449ed506.scope - libcontainer container c05d41a31e0516324b4efaba818e5e7db89659c320c2c67e4323abde449ed506. Jan 21 05:53:01.075000 audit: BPF prog-id=86 op=LOAD Jan 21 05:53:01.076000 audit: BPF prog-id=87 op=LOAD Jan 21 05:53:01.076000 audit[2721]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=2699 pid=2721 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.076000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6332383161613166303061643138326265396235356662366261316530 Jan 21 05:53:01.076000 audit: BPF prog-id=87 op=UNLOAD Jan 21 05:53:01.076000 audit[2721]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2699 pid=2721 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.076000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6332383161613166303061643138326265396235356662366261316530 Jan 21 05:53:01.076000 audit: BPF prog-id=88 op=LOAD Jan 21 05:53:01.076000 audit[2721]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=2699 pid=2721 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.076000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6332383161613166303061643138326265396235356662366261316530 Jan 21 05:53:01.076000 audit: BPF prog-id=89 op=LOAD Jan 21 05:53:01.076000 audit[2721]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=2699 pid=2721 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.076000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6332383161613166303061643138326265396235356662366261316530 Jan 21 05:53:01.076000 audit: BPF prog-id=89 op=UNLOAD Jan 21 05:53:01.076000 audit[2721]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2699 pid=2721 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.076000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6332383161613166303061643138326265396235356662366261316530 Jan 21 05:53:01.077000 audit: BPF prog-id=88 op=UNLOAD Jan 21 05:53:01.077000 audit[2721]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2699 pid=2721 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.077000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6332383161613166303061643138326265396235356662366261316530 Jan 21 05:53:01.077000 audit: BPF prog-id=90 op=LOAD Jan 21 05:53:01.077000 audit[2721]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=2699 pid=2721 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.077000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6332383161613166303061643138326265396235356662366261316530 Jan 21 05:53:01.092089 containerd[1639]: time="2026-01-21T05:53:01.091579288Z" level=info msg="CreateContainer within sandbox \"961c312c87d378139c025716e98fcb3edc7f8ef6c69f561cd7794866bc1be685\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 21 05:53:01.201528 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4230129365.mount: Deactivated successfully. Jan 21 05:53:01.223726 containerd[1639]: time="2026-01-21T05:53:01.222324372Z" level=info msg="Container c3c74bfd5bea7df8442cf52968faed191cd96bb3e663051f306ab2d92af6476f: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:53:01.222532 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1794403611.mount: Deactivated successfully. Jan 21 05:53:01.231288 kubelet[2602]: I0121 05:53:01.231055 2602 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 21 05:53:01.231000 audit: BPF prog-id=91 op=LOAD Jan 21 05:53:01.247000 audit: BPF prog-id=92 op=LOAD Jan 21 05:53:01.247000 audit[2717]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c238 a2=98 a3=0 items=0 ppid=2694 pid=2717 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.247000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6330356434316133316530353136333234623465666162613831386535 Jan 21 05:53:01.248000 audit: BPF prog-id=92 op=UNLOAD Jan 21 05:53:01.248000 audit[2717]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2694 pid=2717 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.248000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6330356434316133316530353136333234623465666162613831386535 Jan 21 05:53:01.249000 audit: BPF prog-id=93 op=LOAD Jan 21 05:53:01.249000 audit[2717]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c488 a2=98 a3=0 items=0 ppid=2694 pid=2717 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.249000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6330356434316133316530353136333234623465666162613831386535 Jan 21 05:53:01.249000 audit: BPF prog-id=94 op=LOAD Jan 21 05:53:01.249000 audit[2717]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00010c218 a2=98 a3=0 items=0 ppid=2694 pid=2717 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.249000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6330356434316133316530353136333234623465666162613831386535 Jan 21 05:53:01.249000 audit: BPF prog-id=94 op=UNLOAD Jan 21 05:53:01.249000 audit[2717]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2694 pid=2717 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.249000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6330356434316133316530353136333234623465666162613831386535 Jan 21 05:53:01.249000 audit: BPF prog-id=93 op=UNLOAD Jan 21 05:53:01.249000 audit[2717]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2694 pid=2717 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.249000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6330356434316133316530353136333234623465666162613831386535 Jan 21 05:53:01.249000 audit: BPF prog-id=95 op=LOAD Jan 21 05:53:01.249000 audit[2717]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c6e8 a2=98 a3=0 items=0 ppid=2694 pid=2717 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.249000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6330356434316133316530353136333234623465666162613831386535 Jan 21 05:53:01.255968 kubelet[2602]: E0121 05:53:01.255281 2602 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.61:6443/api/v1/nodes\": dial tcp 10.0.0.61:6443: connect: connection refused" node="localhost" Jan 21 05:53:01.279821 containerd[1639]: time="2026-01-21T05:53:01.279429139Z" level=info msg="CreateContainer within sandbox \"961c312c87d378139c025716e98fcb3edc7f8ef6c69f561cd7794866bc1be685\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"c3c74bfd5bea7df8442cf52968faed191cd96bb3e663051f306ab2d92af6476f\"" Jan 21 05:53:01.286896 containerd[1639]: time="2026-01-21T05:53:01.286440507Z" level=info msg="StartContainer for \"c3c74bfd5bea7df8442cf52968faed191cd96bb3e663051f306ab2d92af6476f\"" Jan 21 05:53:01.290401 containerd[1639]: time="2026-01-21T05:53:01.289958938Z" level=info msg="connecting to shim c3c74bfd5bea7df8442cf52968faed191cd96bb3e663051f306ab2d92af6476f" address="unix:///run/containerd/s/3a27688d86be83c0433a638d1f3f85749f0c13ee32812c8398e84b0e57b62129" protocol=ttrpc version=3 Jan 21 05:53:01.463937 containerd[1639]: time="2026-01-21T05:53:01.461524927Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6e6cfcfb327385445a9bb0d2bc2fd5d4,Namespace:kube-system,Attempt:0,} returns sandbox id \"c281aa1f00ad182be9b55fb6ba1e0bacf920c343122844348606c6c2ab448bb9\"" Jan 21 05:53:01.468623 kubelet[2602]: E0121 05:53:01.468407 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:01.495113 containerd[1639]: time="2026-01-21T05:53:01.494081602Z" level=info msg="CreateContainer within sandbox \"c281aa1f00ad182be9b55fb6ba1e0bacf920c343122844348606c6c2ab448bb9\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 21 05:53:01.515012 systemd[1]: Started cri-containerd-c3c74bfd5bea7df8442cf52968faed191cd96bb3e663051f306ab2d92af6476f.scope - libcontainer container c3c74bfd5bea7df8442cf52968faed191cd96bb3e663051f306ab2d92af6476f. Jan 21 05:53:01.565287 containerd[1639]: time="2026-01-21T05:53:01.564909862Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:66e26b992bcd7ea6fb75e339cf7a3f7d,Namespace:kube-system,Attempt:0,} returns sandbox id \"c05d41a31e0516324b4efaba818e5e7db89659c320c2c67e4323abde449ed506\"" Jan 21 05:53:01.567543 kubelet[2602]: E0121 05:53:01.567510 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:01.579333 containerd[1639]: time="2026-01-21T05:53:01.579095437Z" level=info msg="Container 850226adb0a3ff93569d12f7017c9bb12aa814c2125204e954714e859838d86a: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:53:01.581943 containerd[1639]: time="2026-01-21T05:53:01.581887031Z" level=info msg="CreateContainer within sandbox \"c05d41a31e0516324b4efaba818e5e7db89659c320c2c67e4323abde449ed506\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 21 05:53:01.584000 audit: BPF prog-id=96 op=LOAD Jan 21 05:53:01.591000 audit: BPF prog-id=97 op=LOAD Jan 21 05:53:01.591000 audit[2774]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2651 pid=2774 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.591000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333633734626664356265613764663834343263663532393638666165 Jan 21 05:53:01.591000 audit: BPF prog-id=97 op=UNLOAD Jan 21 05:53:01.591000 audit[2774]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2651 pid=2774 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.591000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333633734626664356265613764663834343263663532393638666165 Jan 21 05:53:01.591000 audit: BPF prog-id=98 op=LOAD Jan 21 05:53:01.591000 audit[2774]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2651 pid=2774 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.591000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333633734626664356265613764663834343263663532393638666165 Jan 21 05:53:01.591000 audit: BPF prog-id=99 op=LOAD Jan 21 05:53:01.591000 audit[2774]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2651 pid=2774 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.591000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333633734626664356265613764663834343263663532393638666165 Jan 21 05:53:01.591000 audit: BPF prog-id=99 op=UNLOAD Jan 21 05:53:01.591000 audit[2774]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2651 pid=2774 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.591000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333633734626664356265613764663834343263663532393638666165 Jan 21 05:53:01.591000 audit: BPF prog-id=98 op=UNLOAD Jan 21 05:53:01.591000 audit[2774]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2651 pid=2774 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.591000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333633734626664356265613764663834343263663532393638666165 Jan 21 05:53:01.591000 audit: BPF prog-id=100 op=LOAD Jan 21 05:53:01.591000 audit[2774]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2651 pid=2774 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:01.591000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333633734626664356265613764663834343263663532393638666165 Jan 21 05:53:01.653213 containerd[1639]: time="2026-01-21T05:53:01.652833627Z" level=info msg="CreateContainer within sandbox \"c281aa1f00ad182be9b55fb6ba1e0bacf920c343122844348606c6c2ab448bb9\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"850226adb0a3ff93569d12f7017c9bb12aa814c2125204e954714e859838d86a\"" Jan 21 05:53:01.655944 containerd[1639]: time="2026-01-21T05:53:01.654468377Z" level=info msg="StartContainer for \"850226adb0a3ff93569d12f7017c9bb12aa814c2125204e954714e859838d86a\"" Jan 21 05:53:01.659747 containerd[1639]: time="2026-01-21T05:53:01.658895078Z" level=info msg="connecting to shim 850226adb0a3ff93569d12f7017c9bb12aa814c2125204e954714e859838d86a" address="unix:///run/containerd/s/4d9e1d05240c8ee62e2c6248539594665496e95f27db63afb582651c4add99c4" protocol=ttrpc version=3 Jan 21 05:53:01.661094 containerd[1639]: time="2026-01-21T05:53:01.660441360Z" level=info msg="Container e1ed7a8810ce0bf2de2d9f86035ed52d533a50777774d428d2649dc721e1ea13: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:53:01.698800 containerd[1639]: time="2026-01-21T05:53:01.698746480Z" level=info msg="CreateContainer within sandbox \"c05d41a31e0516324b4efaba818e5e7db89659c320c2c67e4323abde449ed506\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"e1ed7a8810ce0bf2de2d9f86035ed52d533a50777774d428d2649dc721e1ea13\"" Jan 21 05:53:01.703362 containerd[1639]: time="2026-01-21T05:53:01.703330610Z" level=info msg="StartContainer for \"e1ed7a8810ce0bf2de2d9f86035ed52d533a50777774d428d2649dc721e1ea13\"" Jan 21 05:53:01.705340 containerd[1639]: time="2026-01-21T05:53:01.705309610Z" level=info msg="connecting to shim e1ed7a8810ce0bf2de2d9f86035ed52d533a50777774d428d2649dc721e1ea13" address="unix:///run/containerd/s/b89edb0b4723cad04b13f0bd2ba5cf81d56f7549127e0bb9b67c48f2bc6de2d3" protocol=ttrpc version=3 Jan 21 05:53:01.982742 systemd[1]: Started cri-containerd-850226adb0a3ff93569d12f7017c9bb12aa814c2125204e954714e859838d86a.scope - libcontainer container 850226adb0a3ff93569d12f7017c9bb12aa814c2125204e954714e859838d86a. Jan 21 05:53:02.087374 systemd[1]: Started cri-containerd-e1ed7a8810ce0bf2de2d9f86035ed52d533a50777774d428d2649dc721e1ea13.scope - libcontainer container e1ed7a8810ce0bf2de2d9f86035ed52d533a50777774d428d2649dc721e1ea13. Jan 21 05:53:02.103000 audit: BPF prog-id=101 op=LOAD Jan 21 05:53:02.104000 audit: BPF prog-id=102 op=LOAD Jan 21 05:53:02.104000 audit[2801]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2699 pid=2801 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.104000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3835303232366164623061336666393335363964313266373031376339 Jan 21 05:53:02.104000 audit: BPF prog-id=102 op=UNLOAD Jan 21 05:53:02.104000 audit[2801]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2699 pid=2801 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.104000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3835303232366164623061336666393335363964313266373031376339 Jan 21 05:53:02.104000 audit: BPF prog-id=103 op=LOAD Jan 21 05:53:02.104000 audit[2801]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2699 pid=2801 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.104000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3835303232366164623061336666393335363964313266373031376339 Jan 21 05:53:02.105000 audit: BPF prog-id=104 op=LOAD Jan 21 05:53:02.105000 audit[2801]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2699 pid=2801 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.105000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3835303232366164623061336666393335363964313266373031376339 Jan 21 05:53:02.105000 audit: BPF prog-id=104 op=UNLOAD Jan 21 05:53:02.105000 audit[2801]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2699 pid=2801 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.105000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3835303232366164623061336666393335363964313266373031376339 Jan 21 05:53:02.105000 audit: BPF prog-id=103 op=UNLOAD Jan 21 05:53:02.105000 audit[2801]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2699 pid=2801 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.105000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3835303232366164623061336666393335363964313266373031376339 Jan 21 05:53:02.105000 audit: BPF prog-id=105 op=LOAD Jan 21 05:53:02.105000 audit[2801]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2699 pid=2801 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.105000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3835303232366164623061336666393335363964313266373031376339 Jan 21 05:53:02.112050 kubelet[2602]: E0121 05:53:02.111961 2602 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.61:6443: connect: connection refused" interval="3.2s" Jan 21 05:53:02.231411 containerd[1639]: time="2026-01-21T05:53:02.230554683Z" level=info msg="StartContainer for \"c3c74bfd5bea7df8442cf52968faed191cd96bb3e663051f306ab2d92af6476f\" returns successfully" Jan 21 05:53:02.258333 kubelet[2602]: E0121 05:53:02.257997 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.61:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 21 05:53:02.301000 audit: BPF prog-id=106 op=LOAD Jan 21 05:53:02.306000 audit: BPF prog-id=107 op=LOAD Jan 21 05:53:02.306000 audit[2812]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=2694 pid=2812 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531656437613838313063653062663264653264396638363033356564 Jan 21 05:53:02.306000 audit: BPF prog-id=107 op=UNLOAD Jan 21 05:53:02.306000 audit[2812]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2694 pid=2812 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531656437613838313063653062663264653264396638363033356564 Jan 21 05:53:02.306000 audit: BPF prog-id=108 op=LOAD Jan 21 05:53:02.306000 audit[2812]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=2694 pid=2812 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531656437613838313063653062663264653264396638363033356564 Jan 21 05:53:02.306000 audit: BPF prog-id=109 op=LOAD Jan 21 05:53:02.306000 audit[2812]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=2694 pid=2812 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531656437613838313063653062663264653264396638363033356564 Jan 21 05:53:02.306000 audit: BPF prog-id=109 op=UNLOAD Jan 21 05:53:02.306000 audit[2812]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2694 pid=2812 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531656437613838313063653062663264653264396638363033356564 Jan 21 05:53:02.306000 audit: BPF prog-id=108 op=UNLOAD Jan 21 05:53:02.306000 audit[2812]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2694 pid=2812 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531656437613838313063653062663264653264396638363033356564 Jan 21 05:53:02.307000 audit: BPF prog-id=110 op=LOAD Jan 21 05:53:02.307000 audit[2812]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=2694 pid=2812 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:02.307000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531656437613838313063653062663264653264396638363033356564 Jan 21 05:53:02.325959 containerd[1639]: time="2026-01-21T05:53:02.325626548Z" level=info msg="StartContainer for \"850226adb0a3ff93569d12f7017c9bb12aa814c2125204e954714e859838d86a\" returns successfully" Jan 21 05:53:02.378779 kubelet[2602]: E0121 05:53:02.378376 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.61:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 21 05:53:02.454335 kubelet[2602]: E0121 05:53:02.454262 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.61:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.61:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 21 05:53:02.501799 containerd[1639]: time="2026-01-21T05:53:02.501520444Z" level=info msg="StartContainer for \"e1ed7a8810ce0bf2de2d9f86035ed52d533a50777774d428d2649dc721e1ea13\" returns successfully" Jan 21 05:53:02.641253 kubelet[2602]: E0121 05:53:02.637899 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:02.644280 kubelet[2602]: E0121 05:53:02.644027 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:02.668298 kubelet[2602]: E0121 05:53:02.668000 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:02.668536 kubelet[2602]: E0121 05:53:02.668425 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:02.675786 kubelet[2602]: E0121 05:53:02.675407 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:02.676272 kubelet[2602]: E0121 05:53:02.676014 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:02.886362 kubelet[2602]: I0121 05:53:02.885921 2602 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 21 05:53:03.691461 kubelet[2602]: E0121 05:53:03.691053 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:03.691461 kubelet[2602]: E0121 05:53:03.692218 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:03.695504 kubelet[2602]: E0121 05:53:03.692950 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:03.695556 kubelet[2602]: E0121 05:53:03.695505 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:04.222331 kubelet[2602]: E0121 05:53:04.222290 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:04.224113 kubelet[2602]: E0121 05:53:04.223919 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:05.102180 kubelet[2602]: E0121 05:53:05.100252 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:05.119781 kubelet[2602]: E0121 05:53:05.106019 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:05.119781 kubelet[2602]: E0121 05:53:05.112498 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:05.130276 kubelet[2602]: E0121 05:53:05.112975 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:06.165033 kubelet[2602]: E0121 05:53:06.163489 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:06.178250 kubelet[2602]: E0121 05:53:06.171440 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:09.584953 kubelet[2602]: E0121 05:53:09.582735 2602 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 21 05:53:11.045798 kubelet[2602]: E0121 05:53:11.042899 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:11.045798 kubelet[2602]: E0121 05:53:11.043265 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:13.112800 kubelet[2602]: E0121 05:53:13.103968 2602 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.61:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 21 05:53:13.112800 kubelet[2602]: E0121 05:53:13.106434 2602 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.61:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Jan 21 05:53:15.426241 kubelet[2602]: E0121 05:53:15.413477 2602 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="6.4s" Jan 21 05:53:15.631540 kubelet[2602]: E0121 05:53:15.604485 2602 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.61:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 21 05:53:16.290868 kubelet[2602]: E0121 05:53:16.279469 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:16.290868 kubelet[2602]: E0121 05:53:16.280913 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:16.325780 kubelet[2602]: I0121 05:53:16.325159 2602 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 21 05:53:16.478068 kubelet[2602]: E0121 05:53:16.474055 2602 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.188ca930658b252a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-21 05:52:58.97690449 +0000 UTC m=+2.710366281,LastTimestamp:2026-01-21 05:52:58.97690449 +0000 UTC m=+2.710366281,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 21 05:53:16.849010 kubelet[2602]: I0121 05:53:16.848504 2602 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jan 21 05:53:16.899497 kubelet[2602]: E0121 05:53:16.849848 2602 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Jan 21 05:53:17.321390 kubelet[2602]: E0121 05:53:17.320708 2602 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 21 05:53:17.421879 kubelet[2602]: E0121 05:53:17.419407 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:17.621941 kubelet[2602]: E0121 05:53:17.614514 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:17.781244 kubelet[2602]: E0121 05:53:17.765539 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:17.901289 kubelet[2602]: E0121 05:53:17.898532 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:18.070434 kubelet[2602]: E0121 05:53:18.066822 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:18.174107 kubelet[2602]: E0121 05:53:18.170226 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:18.274864 kubelet[2602]: E0121 05:53:18.272866 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:18.410195 kubelet[2602]: E0121 05:53:18.407878 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:18.579369 kubelet[2602]: E0121 05:53:18.530351 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:18.672509 kubelet[2602]: E0121 05:53:18.671543 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:18.814434 kubelet[2602]: E0121 05:53:18.812870 2602 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 21 05:53:18.918422 kubelet[2602]: I0121 05:53:18.917064 2602 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 21 05:53:19.306450 kubelet[2602]: I0121 05:53:19.303252 2602 apiserver.go:52] "Watching apiserver" Jan 21 05:53:19.328452 kubelet[2602]: I0121 05:53:19.328393 2602 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jan 21 05:53:19.401060 kubelet[2602]: E0121 05:53:19.399366 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:19.401369 kubelet[2602]: I0121 05:53:19.401290 2602 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jan 21 05:53:19.409739 kubelet[2602]: I0121 05:53:19.407226 2602 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jan 21 05:53:19.417072 kubelet[2602]: E0121 05:53:19.414766 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:19.468715 kubelet[2602]: E0121 05:53:19.468409 2602 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:19.607035 kubelet[2602]: I0121 05:53:19.606482 2602 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=0.606413456 podStartE2EDuration="606.413456ms" podCreationTimestamp="2026-01-21 05:53:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 05:53:19.605799591 +0000 UTC m=+23.339261403" watchObservedRunningTime="2026-01-21 05:53:19.606413456 +0000 UTC m=+23.339875247" Jan 21 05:53:19.803975 kubelet[2602]: I0121 05:53:19.803193 2602 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=0.803170171 podStartE2EDuration="803.170171ms" podCreationTimestamp="2026-01-21 05:53:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 05:53:19.751340353 +0000 UTC m=+23.484802174" watchObservedRunningTime="2026-01-21 05:53:19.803170171 +0000 UTC m=+23.536631972" Jan 21 05:53:19.812700 kubelet[2602]: I0121 05:53:19.812225 2602 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=0.81042888 podStartE2EDuration="810.42888ms" podCreationTimestamp="2026-01-21 05:53:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 05:53:19.810252211 +0000 UTC m=+23.543714043" watchObservedRunningTime="2026-01-21 05:53:19.81042888 +0000 UTC m=+23.543890671" Jan 21 05:53:23.915844 systemd[1]: Reload requested from client PID 2890 ('systemctl') (unit session-10.scope)... Jan 21 05:53:23.915905 systemd[1]: Reloading... Jan 21 05:53:24.169203 zram_generator::config[2942]: No configuration found. Jan 21 05:53:25.038529 systemd[1]: Reloading finished in 1121 ms. Jan 21 05:53:25.129771 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:53:25.151929 systemd[1]: kubelet.service: Deactivated successfully. Jan 21 05:53:25.152496 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:53:25.152000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:53:25.152961 systemd[1]: kubelet.service: Consumed 6.824s CPU time, 132.7M memory peak. Jan 21 05:53:25.157169 kernel: kauditd_printk_skb: 158 callbacks suppressed Jan 21 05:53:25.157349 kernel: audit: type=1131 audit(1768974805.152:396): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:53:25.161869 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 21 05:53:25.161000 audit: BPF prog-id=111 op=LOAD Jan 21 05:53:25.174870 kernel: audit: type=1334 audit(1768974805.161:397): prog-id=111 op=LOAD Jan 21 05:53:25.174950 kernel: audit: type=1334 audit(1768974805.161:398): prog-id=112 op=LOAD Jan 21 05:53:25.161000 audit: BPF prog-id=112 op=LOAD Jan 21 05:53:25.161000 audit: BPF prog-id=68 op=UNLOAD Jan 21 05:53:25.181304 kernel: audit: type=1334 audit(1768974805.161:399): prog-id=68 op=UNLOAD Jan 21 05:53:25.181427 kernel: audit: type=1334 audit(1768974805.161:400): prog-id=69 op=UNLOAD Jan 21 05:53:25.161000 audit: BPF prog-id=69 op=UNLOAD Jan 21 05:53:25.184807 kernel: audit: type=1334 audit(1768974805.170:401): prog-id=113 op=LOAD Jan 21 05:53:25.170000 audit: BPF prog-id=113 op=LOAD Jan 21 05:53:25.170000 audit: BPF prog-id=74 op=UNLOAD Jan 21 05:53:25.170000 audit: BPF prog-id=114 op=LOAD Jan 21 05:53:25.214709 kernel: audit: type=1334 audit(1768974805.170:402): prog-id=74 op=UNLOAD Jan 21 05:53:25.214872 kernel: audit: type=1334 audit(1768974805.170:403): prog-id=114 op=LOAD Jan 21 05:53:25.170000 audit: BPF prog-id=115 op=LOAD Jan 21 05:53:25.170000 audit: BPF prog-id=75 op=UNLOAD Jan 21 05:53:25.224793 kernel: audit: type=1334 audit(1768974805.170:404): prog-id=115 op=LOAD Jan 21 05:53:25.224860 kernel: audit: type=1334 audit(1768974805.170:405): prog-id=75 op=UNLOAD Jan 21 05:53:25.170000 audit: BPF prog-id=76 op=UNLOAD Jan 21 05:53:25.193000 audit: BPF prog-id=116 op=LOAD Jan 21 05:53:25.193000 audit: BPF prog-id=77 op=UNLOAD Jan 21 05:53:25.193000 audit: BPF prog-id=117 op=LOAD Jan 21 05:53:25.193000 audit: BPF prog-id=118 op=LOAD Jan 21 05:53:25.193000 audit: BPF prog-id=78 op=UNLOAD Jan 21 05:53:25.193000 audit: BPF prog-id=79 op=UNLOAD Jan 21 05:53:25.193000 audit: BPF prog-id=119 op=LOAD Jan 21 05:53:25.193000 audit: BPF prog-id=65 op=UNLOAD Jan 21 05:53:25.193000 audit: BPF prog-id=120 op=LOAD Jan 21 05:53:25.193000 audit: BPF prog-id=121 op=LOAD Jan 21 05:53:25.193000 audit: BPF prog-id=66 op=UNLOAD Jan 21 05:53:25.193000 audit: BPF prog-id=67 op=UNLOAD Jan 21 05:53:25.193000 audit: BPF prog-id=122 op=LOAD Jan 21 05:53:25.193000 audit: BPF prog-id=70 op=UNLOAD Jan 21 05:53:25.193000 audit: BPF prog-id=123 op=LOAD Jan 21 05:53:25.193000 audit: BPF prog-id=124 op=LOAD Jan 21 05:53:25.193000 audit: BPF prog-id=71 op=UNLOAD Jan 21 05:53:25.193000 audit: BPF prog-id=72 op=UNLOAD Jan 21 05:53:25.200000 audit: BPF prog-id=125 op=LOAD Jan 21 05:53:25.200000 audit: BPF prog-id=80 op=UNLOAD Jan 21 05:53:25.200000 audit: BPF prog-id=126 op=LOAD Jan 21 05:53:25.200000 audit: BPF prog-id=64 op=UNLOAD Jan 21 05:53:25.205000 audit: BPF prog-id=127 op=LOAD Jan 21 05:53:25.205000 audit: BPF prog-id=73 op=UNLOAD Jan 21 05:53:25.205000 audit: BPF prog-id=128 op=LOAD Jan 21 05:53:25.205000 audit: BPF prog-id=61 op=UNLOAD Jan 21 05:53:25.205000 audit: BPF prog-id=129 op=LOAD Jan 21 05:53:25.205000 audit: BPF prog-id=130 op=LOAD Jan 21 05:53:25.205000 audit: BPF prog-id=62 op=UNLOAD Jan 21 05:53:25.205000 audit: BPF prog-id=63 op=UNLOAD Jan 21 05:53:25.691000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:53:25.691119 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 21 05:53:25.726383 (kubelet)[2982]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 21 05:53:25.895506 kubelet[2982]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 05:53:25.895506 kubelet[2982]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 21 05:53:25.895506 kubelet[2982]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 05:53:25.895506 kubelet[2982]: I0121 05:53:25.895430 2982 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 21 05:53:25.931129 kubelet[2982]: I0121 05:53:25.929032 2982 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jan 21 05:53:25.931129 kubelet[2982]: I0121 05:53:25.929402 2982 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 21 05:53:25.948901 kubelet[2982]: I0121 05:53:25.946137 2982 server.go:956] "Client rotation is on, will bootstrap in background" Jan 21 05:53:25.958995 kubelet[2982]: I0121 05:53:25.957055 2982 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Jan 21 05:53:25.971816 kubelet[2982]: I0121 05:53:25.971734 2982 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 21 05:53:25.986776 kubelet[2982]: I0121 05:53:25.984849 2982 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 21 05:53:26.000722 kubelet[2982]: I0121 05:53:26.000606 2982 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 21 05:53:26.002111 kubelet[2982]: I0121 05:53:26.001943 2982 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 21 05:53:26.002592 kubelet[2982]: I0121 05:53:26.002094 2982 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 21 05:53:26.002592 kubelet[2982]: I0121 05:53:26.002538 2982 topology_manager.go:138] "Creating topology manager with none policy" Jan 21 05:53:26.002592 kubelet[2982]: I0121 05:53:26.002586 2982 container_manager_linux.go:303] "Creating device plugin manager" Jan 21 05:53:26.003344 kubelet[2982]: I0121 05:53:26.002691 2982 state_mem.go:36] "Initialized new in-memory state store" Jan 21 05:53:26.003521 kubelet[2982]: I0121 05:53:26.003484 2982 kubelet.go:480] "Attempting to sync node with API server" Jan 21 05:53:26.003521 kubelet[2982]: I0121 05:53:26.003512 2982 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 21 05:53:26.004147 kubelet[2982]: I0121 05:53:26.003617 2982 kubelet.go:386] "Adding apiserver pod source" Jan 21 05:53:26.004147 kubelet[2982]: I0121 05:53:26.003688 2982 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 21 05:53:26.008963 kubelet[2982]: I0121 05:53:26.008779 2982 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 21 05:53:26.009800 kubelet[2982]: I0121 05:53:26.009479 2982 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 21 05:53:26.037062 kubelet[2982]: I0121 05:53:26.036959 2982 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 21 05:53:26.040902 kubelet[2982]: I0121 05:53:26.037107 2982 server.go:1289] "Started kubelet" Jan 21 05:53:26.055930 kubelet[2982]: I0121 05:53:26.055765 2982 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 21 05:53:26.064280 kubelet[2982]: I0121 05:53:26.064201 2982 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 21 05:53:26.065018 kubelet[2982]: I0121 05:53:26.064897 2982 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 21 05:53:26.083857 kubelet[2982]: I0121 05:53:26.080466 2982 server.go:317] "Adding debug handlers to kubelet server" Jan 21 05:53:26.105522 kubelet[2982]: I0121 05:53:26.103901 2982 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 21 05:53:26.114779 kubelet[2982]: I0121 05:53:26.109471 2982 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 21 05:53:26.124222 kubelet[2982]: I0121 05:53:26.124085 2982 factory.go:223] Registration of the systemd container factory successfully Jan 21 05:53:26.124766 kubelet[2982]: I0121 05:53:26.124257 2982 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 21 05:53:26.125230 kubelet[2982]: I0121 05:53:26.125151 2982 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 21 05:53:26.130961 kubelet[2982]: I0121 05:53:26.129324 2982 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jan 21 05:53:26.138736 kubelet[2982]: E0121 05:53:26.125880 2982 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 21 05:53:26.148940 kubelet[2982]: I0121 05:53:26.141011 2982 reconciler.go:26] "Reconciler: start to sync state" Jan 21 05:53:26.149348 kubelet[2982]: I0121 05:53:26.144523 2982 factory.go:223] Registration of the containerd container factory successfully Jan 21 05:53:26.202972 kubelet[2982]: I0121 05:53:26.201446 2982 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jan 21 05:53:26.207215 kubelet[2982]: I0121 05:53:26.207091 2982 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jan 21 05:53:26.207215 kubelet[2982]: I0121 05:53:26.207151 2982 status_manager.go:230] "Starting to sync pod status with apiserver" Jan 21 05:53:26.207215 kubelet[2982]: I0121 05:53:26.207181 2982 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 21 05:53:26.207215 kubelet[2982]: I0121 05:53:26.207193 2982 kubelet.go:2436] "Starting kubelet main sync loop" Jan 21 05:53:26.207439 kubelet[2982]: E0121 05:53:26.207263 2982 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 21 05:53:26.267911 kubelet[2982]: I0121 05:53:26.267876 2982 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 21 05:53:26.267911 kubelet[2982]: I0121 05:53:26.267903 2982 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 21 05:53:26.268132 kubelet[2982]: I0121 05:53:26.267933 2982 state_mem.go:36] "Initialized new in-memory state store" Jan 21 05:53:26.268173 kubelet[2982]: I0121 05:53:26.268129 2982 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 21 05:53:26.268298 kubelet[2982]: I0121 05:53:26.268191 2982 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 21 05:53:26.268298 kubelet[2982]: I0121 05:53:26.268264 2982 policy_none.go:49] "None policy: Start" Jan 21 05:53:26.268298 kubelet[2982]: I0121 05:53:26.268280 2982 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 21 05:53:26.268298 kubelet[2982]: I0121 05:53:26.268298 2982 state_mem.go:35] "Initializing new in-memory state store" Jan 21 05:53:26.268445 kubelet[2982]: I0121 05:53:26.268429 2982 state_mem.go:75] "Updated machine memory state" Jan 21 05:53:26.296728 kubelet[2982]: E0121 05:53:26.295745 2982 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 21 05:53:26.302448 kubelet[2982]: I0121 05:53:26.298845 2982 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 21 05:53:26.302448 kubelet[2982]: I0121 05:53:26.298989 2982 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 21 05:53:26.302448 kubelet[2982]: I0121 05:53:26.299305 2982 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 21 05:53:26.320461 kubelet[2982]: I0121 05:53:26.313531 2982 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jan 21 05:53:26.320461 kubelet[2982]: I0121 05:53:26.314093 2982 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 21 05:53:26.320461 kubelet[2982]: I0121 05:53:26.314361 2982 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jan 21 05:53:26.320461 kubelet[2982]: E0121 05:53:26.314748 2982 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 21 05:53:26.353468 kubelet[2982]: I0121 05:53:26.352274 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/52e4d30ab66d3d1182c212fc42b84441-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"52e4d30ab66d3d1182c212fc42b84441\") " pod="kube-system/kube-apiserver-localhost" Jan 21 05:53:26.366178 kubelet[2982]: I0121 05:53:26.364813 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:53:26.366178 kubelet[2982]: I0121 05:53:26.364893 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:53:26.366178 kubelet[2982]: I0121 05:53:26.364930 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:53:26.366178 kubelet[2982]: I0121 05:53:26.364956 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:53:26.366178 kubelet[2982]: I0121 05:53:26.364981 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/52e4d30ab66d3d1182c212fc42b84441-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"52e4d30ab66d3d1182c212fc42b84441\") " pod="kube-system/kube-apiserver-localhost" Jan 21 05:53:26.368166 kubelet[2982]: I0121 05:53:26.365010 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 21 05:53:26.368166 kubelet[2982]: I0121 05:53:26.365039 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6e6cfcfb327385445a9bb0d2bc2fd5d4-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6e6cfcfb327385445a9bb0d2bc2fd5d4\") " pod="kube-system/kube-scheduler-localhost" Jan 21 05:53:26.368166 kubelet[2982]: I0121 05:53:26.365065 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/52e4d30ab66d3d1182c212fc42b84441-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"52e4d30ab66d3d1182c212fc42b84441\") " pod="kube-system/kube-apiserver-localhost" Jan 21 05:53:26.398300 kubelet[2982]: E0121 05:53:26.398212 2982 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Jan 21 05:53:26.403731 kubelet[2982]: E0121 05:53:26.403599 2982 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jan 21 05:53:26.408048 kubelet[2982]: E0121 05:53:26.407911 2982 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Jan 21 05:53:26.480889 kubelet[2982]: I0121 05:53:26.479354 2982 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 21 05:53:26.548766 kubelet[2982]: I0121 05:53:26.547615 2982 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Jan 21 05:53:26.548766 kubelet[2982]: I0121 05:53:26.547878 2982 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jan 21 05:53:26.698834 kubelet[2982]: E0121 05:53:26.698780 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:26.707570 kubelet[2982]: E0121 05:53:26.705424 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:26.710037 kubelet[2982]: E0121 05:53:26.709272 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:27.004945 kubelet[2982]: I0121 05:53:27.004908 2982 apiserver.go:52] "Watching apiserver" Jan 21 05:53:27.031444 kubelet[2982]: I0121 05:53:27.031162 2982 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jan 21 05:53:27.248750 kubelet[2982]: E0121 05:53:27.247888 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:27.249170 kubelet[2982]: I0121 05:53:27.248565 2982 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 21 05:53:27.250614 kubelet[2982]: E0121 05:53:27.250203 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:27.284105 kubelet[2982]: E0121 05:53:27.282989 2982 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jan 21 05:53:27.284414 kubelet[2982]: E0121 05:53:27.284329 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:28.076729 kubelet[2982]: I0121 05:53:28.076549 2982 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 21 05:53:28.078379 containerd[1639]: time="2026-01-21T05:53:28.078182942Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 21 05:53:28.078996 kubelet[2982]: I0121 05:53:28.078961 2982 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 21 05:53:28.251694 kubelet[2982]: E0121 05:53:28.251339 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:28.251694 kubelet[2982]: E0121 05:53:28.251411 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:28.905178 kubelet[2982]: E0121 05:53:28.905046 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:29.200327 systemd[1]: Created slice kubepods-besteffort-podb724b754_3e44_438c_b9c3_b1a8b1c65408.slice - libcontainer container kubepods-besteffort-podb724b754_3e44_438c_b9c3_b1a8b1c65408.slice. Jan 21 05:53:29.253715 kubelet[2982]: E0121 05:53:29.253537 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:29.254866 kubelet[2982]: E0121 05:53:29.253989 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:29.312727 kubelet[2982]: I0121 05:53:29.312512 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b724b754-3e44-438c-b9c3-b1a8b1c65408-xtables-lock\") pod \"kube-proxy-jvrww\" (UID: \"b724b754-3e44-438c-b9c3-b1a8b1c65408\") " pod="kube-system/kube-proxy-jvrww" Jan 21 05:53:29.312727 kubelet[2982]: I0121 05:53:29.312602 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/b724b754-3e44-438c-b9c3-b1a8b1c65408-kube-proxy\") pod \"kube-proxy-jvrww\" (UID: \"b724b754-3e44-438c-b9c3-b1a8b1c65408\") " pod="kube-system/kube-proxy-jvrww" Jan 21 05:53:29.314048 kubelet[2982]: I0121 05:53:29.313977 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b724b754-3e44-438c-b9c3-b1a8b1c65408-lib-modules\") pod \"kube-proxy-jvrww\" (UID: \"b724b754-3e44-438c-b9c3-b1a8b1c65408\") " pod="kube-system/kube-proxy-jvrww" Jan 21 05:53:29.314120 kubelet[2982]: I0121 05:53:29.314060 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tgmg\" (UniqueName: \"kubernetes.io/projected/b724b754-3e44-438c-b9c3-b1a8b1c65408-kube-api-access-2tgmg\") pod \"kube-proxy-jvrww\" (UID: \"b724b754-3e44-438c-b9c3-b1a8b1c65408\") " pod="kube-system/kube-proxy-jvrww" Jan 21 05:53:29.361191 systemd[1]: Created slice kubepods-besteffort-pod8421377b_b6da_4b37_abad_787ecdde7ec6.slice - libcontainer container kubepods-besteffort-pod8421377b_b6da_4b37_abad_787ecdde7ec6.slice. Jan 21 05:53:29.415192 kubelet[2982]: I0121 05:53:29.415024 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/8421377b-b6da-4b37-abad-787ecdde7ec6-var-lib-calico\") pod \"tigera-operator-7dcd859c48-8cg2b\" (UID: \"8421377b-b6da-4b37-abad-787ecdde7ec6\") " pod="tigera-operator/tigera-operator-7dcd859c48-8cg2b" Jan 21 05:53:29.415813 kubelet[2982]: I0121 05:53:29.415777 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmztx\" (UniqueName: \"kubernetes.io/projected/8421377b-b6da-4b37-abad-787ecdde7ec6-kube-api-access-wmztx\") pod \"tigera-operator-7dcd859c48-8cg2b\" (UID: \"8421377b-b6da-4b37-abad-787ecdde7ec6\") " pod="tigera-operator/tigera-operator-7dcd859c48-8cg2b" Jan 21 05:53:29.521738 kubelet[2982]: E0121 05:53:29.521465 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:29.525362 containerd[1639]: time="2026-01-21T05:53:29.525004739Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jvrww,Uid:b724b754-3e44-438c-b9c3-b1a8b1c65408,Namespace:kube-system,Attempt:0,}" Jan 21 05:53:29.661794 containerd[1639]: time="2026-01-21T05:53:29.661595738Z" level=info msg="connecting to shim 9a1a06e88ac340ec989bec5ee665275413e7661bd508681509e04fefca4a21bc" address="unix:///run/containerd/s/d3dbfd4dd21fc5172378f444ce9deaad5258d477f5a2711f12de778b37f083ea" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:53:29.674022 containerd[1639]: time="2026-01-21T05:53:29.673487601Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-8cg2b,Uid:8421377b-b6da-4b37-abad-787ecdde7ec6,Namespace:tigera-operator,Attempt:0,}" Jan 21 05:53:29.769471 containerd[1639]: time="2026-01-21T05:53:29.769327399Z" level=info msg="connecting to shim d1109cd7046ca61851fd487ad12198e93e01fa80a5fc05396c0abf6328ab473f" address="unix:///run/containerd/s/d1f2c581171c562c3bd3057b3e8d7a227d052de2907e8190e6e1436a3e1f2fb2" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:53:29.865099 systemd[1]: Started cri-containerd-9a1a06e88ac340ec989bec5ee665275413e7661bd508681509e04fefca4a21bc.scope - libcontainer container 9a1a06e88ac340ec989bec5ee665275413e7661bd508681509e04fefca4a21bc. Jan 21 05:53:29.896331 systemd[1]: Started cri-containerd-d1109cd7046ca61851fd487ad12198e93e01fa80a5fc05396c0abf6328ab473f.scope - libcontainer container d1109cd7046ca61851fd487ad12198e93e01fa80a5fc05396c0abf6328ab473f. Jan 21 05:53:29.905000 audit: BPF prog-id=131 op=LOAD Jan 21 05:53:29.908000 audit: BPF prog-id=132 op=LOAD Jan 21 05:53:29.908000 audit[3059]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=3048 pid=3059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.908000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961316130366538386163333430656339383962656335656536363532 Jan 21 05:53:29.909000 audit: BPF prog-id=132 op=UNLOAD Jan 21 05:53:29.909000 audit[3059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3048 pid=3059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.909000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961316130366538386163333430656339383962656335656536363532 Jan 21 05:53:29.910000 audit: BPF prog-id=133 op=LOAD Jan 21 05:53:29.910000 audit[3059]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3048 pid=3059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.910000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961316130366538386163333430656339383962656335656536363532 Jan 21 05:53:29.910000 audit: BPF prog-id=134 op=LOAD Jan 21 05:53:29.910000 audit[3059]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3048 pid=3059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.910000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961316130366538386163333430656339383962656335656536363532 Jan 21 05:53:29.910000 audit: BPF prog-id=134 op=UNLOAD Jan 21 05:53:29.910000 audit[3059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3048 pid=3059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.910000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961316130366538386163333430656339383962656335656536363532 Jan 21 05:53:29.910000 audit: BPF prog-id=133 op=UNLOAD Jan 21 05:53:29.910000 audit[3059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3048 pid=3059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.910000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961316130366538386163333430656339383962656335656536363532 Jan 21 05:53:29.910000 audit: BPF prog-id=135 op=LOAD Jan 21 05:53:29.910000 audit[3059]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=3048 pid=3059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.910000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961316130366538386163333430656339383962656335656536363532 Jan 21 05:53:29.945000 audit: BPF prog-id=136 op=LOAD Jan 21 05:53:29.954000 audit: BPF prog-id=137 op=LOAD Jan 21 05:53:29.954000 audit[3086]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00016e238 a2=98 a3=0 items=0 ppid=3069 pid=3086 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.954000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431313039636437303436636136313835316664343837616431323139 Jan 21 05:53:29.956000 audit: BPF prog-id=137 op=UNLOAD Jan 21 05:53:29.956000 audit[3086]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3069 pid=3086 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.956000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431313039636437303436636136313835316664343837616431323139 Jan 21 05:53:29.957000 audit: BPF prog-id=138 op=LOAD Jan 21 05:53:29.957000 audit[3086]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00016e488 a2=98 a3=0 items=0 ppid=3069 pid=3086 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.957000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431313039636437303436636136313835316664343837616431323139 Jan 21 05:53:29.961000 audit: BPF prog-id=139 op=LOAD Jan 21 05:53:29.961000 audit[3086]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00016e218 a2=98 a3=0 items=0 ppid=3069 pid=3086 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.961000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431313039636437303436636136313835316664343837616431323139 Jan 21 05:53:29.962000 audit: BPF prog-id=139 op=UNLOAD Jan 21 05:53:29.962000 audit[3086]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3069 pid=3086 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.962000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431313039636437303436636136313835316664343837616431323139 Jan 21 05:53:29.963000 audit: BPF prog-id=138 op=UNLOAD Jan 21 05:53:29.963000 audit[3086]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3069 pid=3086 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.963000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431313039636437303436636136313835316664343837616431323139 Jan 21 05:53:29.973000 audit: BPF prog-id=140 op=LOAD Jan 21 05:53:29.973000 audit[3086]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00016e6e8 a2=98 a3=0 items=0 ppid=3069 pid=3086 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:29.973000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431313039636437303436636136313835316664343837616431323139 Jan 21 05:53:29.986281 containerd[1639]: time="2026-01-21T05:53:29.985095485Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jvrww,Uid:b724b754-3e44-438c-b9c3-b1a8b1c65408,Namespace:kube-system,Attempt:0,} returns sandbox id \"9a1a06e88ac340ec989bec5ee665275413e7661bd508681509e04fefca4a21bc\"" Jan 21 05:53:29.986950 kubelet[2982]: E0121 05:53:29.986864 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:30.020570 containerd[1639]: time="2026-01-21T05:53:30.018885272Z" level=info msg="CreateContainer within sandbox \"9a1a06e88ac340ec989bec5ee665275413e7661bd508681509e04fefca4a21bc\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 21 05:53:30.065400 containerd[1639]: time="2026-01-21T05:53:30.065225826Z" level=info msg="Container d52352c33020e20e767ed9ae5b1e2c6c9796c0d8a43e690388925b5039641052: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:53:30.075823 containerd[1639]: time="2026-01-21T05:53:30.075767870Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-8cg2b,Uid:8421377b-b6da-4b37-abad-787ecdde7ec6,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"d1109cd7046ca61851fd487ad12198e93e01fa80a5fc05396c0abf6328ab473f\"" Jan 21 05:53:30.080147 containerd[1639]: time="2026-01-21T05:53:30.080051950Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Jan 21 05:53:30.100012 containerd[1639]: time="2026-01-21T05:53:30.099916576Z" level=info msg="CreateContainer within sandbox \"9a1a06e88ac340ec989bec5ee665275413e7661bd508681509e04fefca4a21bc\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"d52352c33020e20e767ed9ae5b1e2c6c9796c0d8a43e690388925b5039641052\"" Jan 21 05:53:30.103710 containerd[1639]: time="2026-01-21T05:53:30.103450606Z" level=info msg="StartContainer for \"d52352c33020e20e767ed9ae5b1e2c6c9796c0d8a43e690388925b5039641052\"" Jan 21 05:53:30.118132 containerd[1639]: time="2026-01-21T05:53:30.116419118Z" level=info msg="connecting to shim d52352c33020e20e767ed9ae5b1e2c6c9796c0d8a43e690388925b5039641052" address="unix:///run/containerd/s/d3dbfd4dd21fc5172378f444ce9deaad5258d477f5a2711f12de778b37f083ea" protocol=ttrpc version=3 Jan 21 05:53:30.189604 systemd[1]: Started cri-containerd-d52352c33020e20e767ed9ae5b1e2c6c9796c0d8a43e690388925b5039641052.scope - libcontainer container d52352c33020e20e767ed9ae5b1e2c6c9796c0d8a43e690388925b5039641052. Jan 21 05:53:30.264244 kubelet[2982]: E0121 05:53:30.264022 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:30.340000 audit: BPF prog-id=141 op=LOAD Jan 21 05:53:30.345296 kernel: kauditd_printk_skb: 76 callbacks suppressed Jan 21 05:53:30.345482 kernel: audit: type=1334 audit(1768974810.340:454): prog-id=141 op=LOAD Jan 21 05:53:30.340000 audit[3131]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3048 pid=3131 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.366291 kernel: audit: type=1300 audit(1768974810.340:454): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3048 pid=3131 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.340000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6435323335326333333032306532306537363765643961653562316532 Jan 21 05:53:30.340000 audit: BPF prog-id=142 op=LOAD Jan 21 05:53:30.390758 kernel: audit: type=1327 audit(1768974810.340:454): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6435323335326333333032306532306537363765643961653562316532 Jan 21 05:53:30.390843 kernel: audit: type=1334 audit(1768974810.340:455): prog-id=142 op=LOAD Jan 21 05:53:30.340000 audit[3131]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3048 pid=3131 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.340000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6435323335326333333032306532306537363765643961653562316532 Jan 21 05:53:30.423575 kernel: audit: type=1300 audit(1768974810.340:455): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3048 pid=3131 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.423799 kernel: audit: type=1327 audit(1768974810.340:455): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6435323335326333333032306532306537363765643961653562316532 Jan 21 05:53:30.423843 kernel: audit: type=1334 audit(1768974810.340:456): prog-id=142 op=UNLOAD Jan 21 05:53:30.340000 audit: BPF prog-id=142 op=UNLOAD Jan 21 05:53:30.340000 audit[3131]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3048 pid=3131 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.444199 kernel: audit: type=1300 audit(1768974810.340:456): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3048 pid=3131 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.446129 kernel: audit: type=1327 audit(1768974810.340:456): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6435323335326333333032306532306537363765643961653562316532 Jan 21 05:53:30.340000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6435323335326333333032306532306537363765643961653562316532 Jan 21 05:53:30.457831 containerd[1639]: time="2026-01-21T05:53:30.457781845Z" level=info msg="StartContainer for \"d52352c33020e20e767ed9ae5b1e2c6c9796c0d8a43e690388925b5039641052\" returns successfully" Jan 21 05:53:30.461870 kernel: audit: type=1334 audit(1768974810.340:457): prog-id=141 op=UNLOAD Jan 21 05:53:30.340000 audit: BPF prog-id=141 op=UNLOAD Jan 21 05:53:30.340000 audit[3131]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3048 pid=3131 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.340000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6435323335326333333032306532306537363765643961653562316532 Jan 21 05:53:30.340000 audit: BPF prog-id=143 op=LOAD Jan 21 05:53:30.340000 audit[3131]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3048 pid=3131 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.340000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6435323335326333333032306532306537363765643961653562316532 Jan 21 05:53:30.860000 audit[3199]: NETFILTER_CFG table=mangle:54 family=2 entries=1 op=nft_register_chain pid=3199 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:30.860000 audit[3199]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe0dfdf1d0 a2=0 a3=7ffe0dfdf1bc items=0 ppid=3144 pid=3199 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.860000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 21 05:53:30.861000 audit[3200]: NETFILTER_CFG table=mangle:55 family=10 entries=1 op=nft_register_chain pid=3200 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:30.861000 audit[3200]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff4bdd5f70 a2=0 a3=7fff4bdd5f5c items=0 ppid=3144 pid=3200 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.861000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 21 05:53:30.870000 audit[3202]: NETFILTER_CFG table=nat:56 family=10 entries=1 op=nft_register_chain pid=3202 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:30.870000 audit[3202]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fffe768f650 a2=0 a3=7fffe768f63c items=0 ppid=3144 pid=3202 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.870000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 21 05:53:30.874000 audit[3205]: NETFILTER_CFG table=nat:57 family=2 entries=1 op=nft_register_chain pid=3205 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:30.874000 audit[3205]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc886cb160 a2=0 a3=7ffc886cb14c items=0 ppid=3144 pid=3205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.874000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 21 05:53:30.880000 audit[3207]: NETFILTER_CFG table=filter:58 family=10 entries=1 op=nft_register_chain pid=3207 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:30.880000 audit[3207]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffdbd6d5930 a2=0 a3=7ffdbd6d591c items=0 ppid=3144 pid=3207 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.880000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 21 05:53:30.881000 audit[3209]: NETFILTER_CFG table=filter:59 family=2 entries=1 op=nft_register_chain pid=3209 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:30.881000 audit[3209]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe2ac21300 a2=0 a3=7ffe2ac212ec items=0 ppid=3144 pid=3209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.881000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 21 05:53:30.971000 audit[3210]: NETFILTER_CFG table=filter:60 family=2 entries=1 op=nft_register_chain pid=3210 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:30.971000 audit[3210]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffee25c24c0 a2=0 a3=7ffee25c24ac items=0 ppid=3144 pid=3210 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.971000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 21 05:53:30.983000 audit[3212]: NETFILTER_CFG table=filter:61 family=2 entries=1 op=nft_register_rule pid=3212 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:30.983000 audit[3212]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffeb9a56f10 a2=0 a3=7ffeb9a56efc items=0 ppid=3144 pid=3212 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:30.983000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276696365 Jan 21 05:53:31.000000 audit[3215]: NETFILTER_CFG table=filter:62 family=2 entries=1 op=nft_register_rule pid=3215 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.000000 audit[3215]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffce26d0020 a2=0 a3=7ffce26d000c items=0 ppid=3144 pid=3215 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.000000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669 Jan 21 05:53:31.005000 audit[3216]: NETFILTER_CFG table=filter:63 family=2 entries=1 op=nft_register_chain pid=3216 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.005000 audit[3216]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe2cb27be0 a2=0 a3=7ffe2cb27bcc items=0 ppid=3144 pid=3216 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.005000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 21 05:53:31.012000 audit[3218]: NETFILTER_CFG table=filter:64 family=2 entries=1 op=nft_register_rule pid=3218 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.012000 audit[3218]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffceca70d00 a2=0 a3=7ffceca70cec items=0 ppid=3144 pid=3218 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.012000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 21 05:53:31.018000 audit[3219]: NETFILTER_CFG table=filter:65 family=2 entries=1 op=nft_register_chain pid=3219 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.018000 audit[3219]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe02948210 a2=0 a3=7ffe029481fc items=0 ppid=3144 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.018000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Jan 21 05:53:31.034000 audit[3221]: NETFILTER_CFG table=filter:66 family=2 entries=1 op=nft_register_rule pid=3221 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.034000 audit[3221]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffdad2b1da0 a2=0 a3=7ffdad2b1d8c items=0 ppid=3144 pid=3221 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.034000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Jan 21 05:53:31.053000 audit[3224]: NETFILTER_CFG table=filter:67 family=2 entries=1 op=nft_register_rule pid=3224 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.053000 audit[3224]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7fffb0df6530 a2=0 a3=7fffb0df651c items=0 ppid=3144 pid=3224 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.053000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D53 Jan 21 05:53:31.057000 audit[3225]: NETFILTER_CFG table=filter:68 family=2 entries=1 op=nft_register_chain pid=3225 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.057000 audit[3225]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc5ee9a250 a2=0 a3=7ffc5ee9a23c items=0 ppid=3144 pid=3225 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.057000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Jan 21 05:53:31.065000 audit[3227]: NETFILTER_CFG table=filter:69 family=2 entries=1 op=nft_register_rule pid=3227 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.065000 audit[3227]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7fff2546a830 a2=0 a3=7fff2546a81c items=0 ppid=3144 pid=3227 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.065000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 21 05:53:31.069000 audit[3228]: NETFILTER_CFG table=filter:70 family=2 entries=1 op=nft_register_chain pid=3228 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.069000 audit[3228]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffcdcce6cd0 a2=0 a3=7ffcdcce6cbc items=0 ppid=3144 pid=3228 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.069000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 21 05:53:31.079000 audit[3230]: NETFILTER_CFG table=filter:71 family=2 entries=1 op=nft_register_rule pid=3230 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.079000 audit[3230]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc1df1b3f0 a2=0 a3=7ffc1df1b3dc items=0 ppid=3144 pid=3230 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.079000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Jan 21 05:53:31.092000 audit[3233]: NETFILTER_CFG table=filter:72 family=2 entries=1 op=nft_register_rule pid=3233 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.092000 audit[3233]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffe68362920 a2=0 a3=7ffe6836290c items=0 ppid=3144 pid=3233 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.092000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Jan 21 05:53:31.106000 audit[3236]: NETFILTER_CFG table=filter:73 family=2 entries=1 op=nft_register_rule pid=3236 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.106000 audit[3236]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7fff88eeb250 a2=0 a3=7fff88eeb23c items=0 ppid=3144 pid=3236 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.106000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Jan 21 05:53:31.112000 audit[3237]: NETFILTER_CFG table=nat:74 family=2 entries=1 op=nft_register_chain pid=3237 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.112000 audit[3237]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe52d10fb0 a2=0 a3=7ffe52d10f9c items=0 ppid=3144 pid=3237 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.112000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Jan 21 05:53:31.128000 audit[3239]: NETFILTER_CFG table=nat:75 family=2 entries=1 op=nft_register_rule pid=3239 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.128000 audit[3239]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7fff56a36df0 a2=0 a3=7fff56a36ddc items=0 ppid=3144 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.128000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 21 05:53:31.145000 audit[3242]: NETFILTER_CFG table=nat:76 family=2 entries=1 op=nft_register_rule pid=3242 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.145000 audit[3242]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7fff69933080 a2=0 a3=7fff6993306c items=0 ppid=3144 pid=3242 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.145000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 21 05:53:31.148000 audit[3243]: NETFILTER_CFG table=nat:77 family=2 entries=1 op=nft_register_chain pid=3243 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.148000 audit[3243]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff125146d0 a2=0 a3=7fff125146bc items=0 ppid=3144 pid=3243 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.148000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 21 05:53:31.159000 audit[3245]: NETFILTER_CFG table=nat:78 family=2 entries=1 op=nft_register_rule pid=3245 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 21 05:53:31.159000 audit[3245]: SYSCALL arch=c000003e syscall=46 success=yes exit=532 a0=3 a1=7ffd27279be0 a2=0 a3=7ffd27279bcc items=0 ppid=3144 pid=3245 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.159000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 21 05:53:31.233000 audit[3251]: NETFILTER_CFG table=filter:79 family=2 entries=8 op=nft_register_rule pid=3251 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:53:31.233000 audit[3251]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffc9f16b340 a2=0 a3=7ffc9f16b32c items=0 ppid=3144 pid=3251 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.233000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:53:31.252000 audit[3251]: NETFILTER_CFG table=nat:80 family=2 entries=14 op=nft_register_chain pid=3251 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:53:31.252000 audit[3251]: SYSCALL arch=c000003e syscall=46 success=yes exit=5508 a0=3 a1=7ffc9f16b340 a2=0 a3=7ffc9f16b32c items=0 ppid=3144 pid=3251 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.252000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:53:31.259000 audit[3256]: NETFILTER_CFG table=filter:81 family=10 entries=1 op=nft_register_chain pid=3256 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.259000 audit[3256]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffdcf51dde0 a2=0 a3=7ffdcf51ddcc items=0 ppid=3144 pid=3256 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.259000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 21 05:53:31.271000 audit[3258]: NETFILTER_CFG table=filter:82 family=10 entries=2 op=nft_register_chain pid=3258 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.271000 audit[3258]: SYSCALL arch=c000003e syscall=46 success=yes exit=836 a0=3 a1=7ffe6fbe9520 a2=0 a3=7ffe6fbe950c items=0 ppid=3144 pid=3258 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.271000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C6520736572766963 Jan 21 05:53:31.273737 kubelet[2982]: E0121 05:53:31.273602 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:31.299000 audit[3261]: NETFILTER_CFG table=filter:83 family=10 entries=1 op=nft_register_rule pid=3261 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.299000 audit[3261]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffda53776f0 a2=0 a3=7ffda53776dc items=0 ppid=3144 pid=3261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.299000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276 Jan 21 05:53:31.305000 audit[3262]: NETFILTER_CFG table=filter:84 family=10 entries=1 op=nft_register_chain pid=3262 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.305000 audit[3262]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcab351a90 a2=0 a3=7ffcab351a7c items=0 ppid=3144 pid=3262 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.305000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 21 05:53:31.317000 audit[3264]: NETFILTER_CFG table=filter:85 family=10 entries=1 op=nft_register_rule pid=3264 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.317000 audit[3264]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffda12b7900 a2=0 a3=7ffda12b78ec items=0 ppid=3144 pid=3264 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.317000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 21 05:53:31.330000 audit[3265]: NETFILTER_CFG table=filter:86 family=10 entries=1 op=nft_register_chain pid=3265 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.330000 audit[3265]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff68a2c3e0 a2=0 a3=7fff68a2c3cc items=0 ppid=3144 pid=3265 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.330000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Jan 21 05:53:31.344000 audit[3267]: NETFILTER_CFG table=filter:87 family=10 entries=1 op=nft_register_rule pid=3267 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.344000 audit[3267]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffc69b30b30 a2=0 a3=7ffc69b30b1c items=0 ppid=3144 pid=3267 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.344000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B554245 Jan 21 05:53:31.366000 audit[3270]: NETFILTER_CFG table=filter:88 family=10 entries=2 op=nft_register_chain pid=3270 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.366000 audit[3270]: SYSCALL arch=c000003e syscall=46 success=yes exit=828 a0=3 a1=7fffa2bb9400 a2=0 a3=7fffa2bb93ec items=0 ppid=3144 pid=3270 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.366000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Jan 21 05:53:31.377000 audit[3271]: NETFILTER_CFG table=filter:89 family=10 entries=1 op=nft_register_chain pid=3271 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.377000 audit[3271]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff7577d0c0 a2=0 a3=7fff7577d0ac items=0 ppid=3144 pid=3271 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.377000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Jan 21 05:53:31.389000 audit[3273]: NETFILTER_CFG table=filter:90 family=10 entries=1 op=nft_register_rule pid=3273 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.389000 audit[3273]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffd4906bfa0 a2=0 a3=7ffd4906bf8c items=0 ppid=3144 pid=3273 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.389000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 21 05:53:31.395000 audit[3274]: NETFILTER_CFG table=filter:91 family=10 entries=1 op=nft_register_chain pid=3274 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.395000 audit[3274]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd6d7c9cd0 a2=0 a3=7ffd6d7c9cbc items=0 ppid=3144 pid=3274 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.395000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 21 05:53:31.410000 audit[3276]: NETFILTER_CFG table=filter:92 family=10 entries=1 op=nft_register_rule pid=3276 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.410000 audit[3276]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc2c4def40 a2=0 a3=7ffc2c4def2c items=0 ppid=3144 pid=3276 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.410000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Jan 21 05:53:31.429000 audit[3279]: NETFILTER_CFG table=filter:93 family=10 entries=1 op=nft_register_rule pid=3279 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.429000 audit[3279]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc10149810 a2=0 a3=7ffc101497fc items=0 ppid=3144 pid=3279 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.429000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Jan 21 05:53:31.448000 audit[3282]: NETFILTER_CFG table=filter:94 family=10 entries=1 op=nft_register_rule pid=3282 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.448000 audit[3282]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffedf205410 a2=0 a3=7ffedf2053fc items=0 ppid=3144 pid=3282 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.448000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C Jan 21 05:53:31.451000 audit[3283]: NETFILTER_CFG table=nat:95 family=10 entries=1 op=nft_register_chain pid=3283 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.451000 audit[3283]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffd61fb1820 a2=0 a3=7ffd61fb180c items=0 ppid=3144 pid=3283 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.451000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Jan 21 05:53:31.477000 audit[3285]: NETFILTER_CFG table=nat:96 family=10 entries=1 op=nft_register_rule pid=3285 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.477000 audit[3285]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffc7cba3910 a2=0 a3=7ffc7cba38fc items=0 ppid=3144 pid=3285 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.477000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 21 05:53:31.494000 audit[3288]: NETFILTER_CFG table=nat:97 family=10 entries=1 op=nft_register_rule pid=3288 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.494000 audit[3288]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7fff0b66ac50 a2=0 a3=7fff0b66ac3c items=0 ppid=3144 pid=3288 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.494000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 21 05:53:31.502000 audit[3289]: NETFILTER_CFG table=nat:98 family=10 entries=1 op=nft_register_chain pid=3289 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.502000 audit[3289]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd6d278070 a2=0 a3=7ffd6d27805c items=0 ppid=3144 pid=3289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.502000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 21 05:53:31.514000 audit[3291]: NETFILTER_CFG table=nat:99 family=10 entries=2 op=nft_register_chain pid=3291 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.514000 audit[3291]: SYSCALL arch=c000003e syscall=46 success=yes exit=612 a0=3 a1=7ffd56865ee0 a2=0 a3=7ffd56865ecc items=0 ppid=3144 pid=3291 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.514000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 21 05:53:31.521000 audit[3292]: NETFILTER_CFG table=filter:100 family=10 entries=1 op=nft_register_chain pid=3292 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.521000 audit[3292]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe473f01a0 a2=0 a3=7ffe473f018c items=0 ppid=3144 pid=3292 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.521000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 21 05:53:31.534000 audit[3294]: NETFILTER_CFG table=filter:101 family=10 entries=1 op=nft_register_rule pid=3294 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.534000 audit[3294]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7fffc699a510 a2=0 a3=7fffc699a4fc items=0 ppid=3144 pid=3294 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.534000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 21 05:53:31.558000 audit[3297]: NETFILTER_CFG table=filter:102 family=10 entries=1 op=nft_register_rule pid=3297 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 21 05:53:31.558000 audit[3297]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffdab02ad60 a2=0 a3=7ffdab02ad4c items=0 ppid=3144 pid=3297 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.558000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 21 05:53:31.582000 audit[3299]: NETFILTER_CFG table=filter:103 family=10 entries=3 op=nft_register_rule pid=3299 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 21 05:53:31.582000 audit[3299]: SYSCALL arch=c000003e syscall=46 success=yes exit=2088 a0=3 a1=7ffd42091570 a2=0 a3=7ffd4209155c items=0 ppid=3144 pid=3299 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.582000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:53:31.583000 audit[3299]: NETFILTER_CFG table=nat:104 family=10 entries=7 op=nft_register_chain pid=3299 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 21 05:53:31.583000 audit[3299]: SYSCALL arch=c000003e syscall=46 success=yes exit=2056 a0=3 a1=7ffd42091570 a2=0 a3=7ffd4209155c items=0 ppid=3144 pid=3299 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:31.583000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:53:31.993727 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1412407750.mount: Deactivated successfully. Jan 21 05:53:32.285099 kubelet[2982]: E0121 05:53:32.284831 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:33.834348 kubelet[2982]: E0121 05:53:33.834142 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:33.882735 kubelet[2982]: I0121 05:53:33.882496 2982 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-jvrww" podStartSLOduration=4.8824724 podStartE2EDuration="4.8824724s" podCreationTimestamp="2026-01-21 05:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 05:53:31.322462139 +0000 UTC m=+5.532111862" watchObservedRunningTime="2026-01-21 05:53:33.8824724 +0000 UTC m=+8.092122104" Jan 21 05:53:34.289985 kubelet[2982]: E0121 05:53:34.289869 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:53:34.451362 containerd[1639]: time="2026-01-21T05:53:34.448469489Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:53:34.461908 containerd[1639]: time="2026-01-21T05:53:34.458573220Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=23558205" Jan 21 05:53:34.474389 containerd[1639]: time="2026-01-21T05:53:34.470608512Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:53:34.478966 containerd[1639]: time="2026-01-21T05:53:34.477466422Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:53:34.478966 containerd[1639]: time="2026-01-21T05:53:34.478520634Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 4.398380966s" Jan 21 05:53:34.478966 containerd[1639]: time="2026-01-21T05:53:34.478554855Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Jan 21 05:53:34.563275 containerd[1639]: time="2026-01-21T05:53:34.551099390Z" level=info msg="CreateContainer within sandbox \"d1109cd7046ca61851fd487ad12198e93e01fa80a5fc05396c0abf6328ab473f\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jan 21 05:53:34.690849 containerd[1639]: time="2026-01-21T05:53:34.688332435Z" level=info msg="Container fc02db2082ca6418649bc848b59c3d10d4f482bf6563db3c653dc87298c8d8cf: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:53:34.751785 containerd[1639]: time="2026-01-21T05:53:34.751470734Z" level=info msg="CreateContainer within sandbox \"d1109cd7046ca61851fd487ad12198e93e01fa80a5fc05396c0abf6328ab473f\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"fc02db2082ca6418649bc848b59c3d10d4f482bf6563db3c653dc87298c8d8cf\"" Jan 21 05:53:34.762244 containerd[1639]: time="2026-01-21T05:53:34.759453913Z" level=info msg="StartContainer for \"fc02db2082ca6418649bc848b59c3d10d4f482bf6563db3c653dc87298c8d8cf\"" Jan 21 05:53:34.764013 containerd[1639]: time="2026-01-21T05:53:34.763972272Z" level=info msg="connecting to shim fc02db2082ca6418649bc848b59c3d10d4f482bf6563db3c653dc87298c8d8cf" address="unix:///run/containerd/s/d1f2c581171c562c3bd3057b3e8d7a227d052de2907e8190e6e1436a3e1f2fb2" protocol=ttrpc version=3 Jan 21 05:53:34.891755 systemd[1]: Started cri-containerd-fc02db2082ca6418649bc848b59c3d10d4f482bf6563db3c653dc87298c8d8cf.scope - libcontainer container fc02db2082ca6418649bc848b59c3d10d4f482bf6563db3c653dc87298c8d8cf. Jan 21 05:53:35.002000 audit: BPF prog-id=144 op=LOAD Jan 21 05:53:35.004000 audit: BPF prog-id=145 op=LOAD Jan 21 05:53:35.004000 audit[3308]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3069 pid=3308 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:35.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663303264623230383263613634313836343962633834386235396333 Jan 21 05:53:35.004000 audit: BPF prog-id=145 op=UNLOAD Jan 21 05:53:35.004000 audit[3308]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3069 pid=3308 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:35.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663303264623230383263613634313836343962633834386235396333 Jan 21 05:53:35.004000 audit: BPF prog-id=146 op=LOAD Jan 21 05:53:35.004000 audit[3308]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3069 pid=3308 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:35.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663303264623230383263613634313836343962633834386235396333 Jan 21 05:53:35.004000 audit: BPF prog-id=147 op=LOAD Jan 21 05:53:35.004000 audit[3308]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3069 pid=3308 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:35.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663303264623230383263613634313836343962633834386235396333 Jan 21 05:53:35.004000 audit: BPF prog-id=147 op=UNLOAD Jan 21 05:53:35.004000 audit[3308]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3069 pid=3308 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:35.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663303264623230383263613634313836343962633834386235396333 Jan 21 05:53:35.005000 audit: BPF prog-id=146 op=UNLOAD Jan 21 05:53:35.005000 audit[3308]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3069 pid=3308 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:35.005000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663303264623230383263613634313836343962633834386235396333 Jan 21 05:53:35.005000 audit: BPF prog-id=148 op=LOAD Jan 21 05:53:35.005000 audit[3308]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3069 pid=3308 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:53:35.005000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663303264623230383263613634313836343962633834386235396333 Jan 21 05:53:35.153603 containerd[1639]: time="2026-01-21T05:53:35.153033875Z" level=info msg="StartContainer for \"fc02db2082ca6418649bc848b59c3d10d4f482bf6563db3c653dc87298c8d8cf\" returns successfully" Jan 21 05:53:35.474091 kubelet[2982]: I0121 05:53:35.468911 2982 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-8cg2b" podStartSLOduration=2.066265573 podStartE2EDuration="6.468885353s" podCreationTimestamp="2026-01-21 05:53:29 +0000 UTC" firstStartedPulling="2026-01-21 05:53:30.079183693 +0000 UTC m=+4.288833406" lastFinishedPulling="2026-01-21 05:53:34.481803484 +0000 UTC m=+8.691453186" observedRunningTime="2026-01-21 05:53:35.468565871 +0000 UTC m=+9.678215574" watchObservedRunningTime="2026-01-21 05:53:35.468885353 +0000 UTC m=+9.678535055" Jan 21 05:53:41.178334 kubelet[2982]: E0121 05:53:41.178140 2982 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.755s" Jan 21 05:53:52.663000 audit[1880]: USER_END pid=1880 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:53:52.673130 kernel: kauditd_printk_skb: 180 callbacks suppressed Jan 21 05:53:52.673405 kernel: audit: type=1106 audit(1768974832.663:518): pid=1880 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:53:52.663855 sudo[1880]: pam_unix(sudo:session): session closed for user root Jan 21 05:53:52.663000 audit[1880]: CRED_DISP pid=1880 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:53:52.700629 kernel: audit: type=1104 audit(1768974832.663:519): pid=1880 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 21 05:53:52.701499 sshd[1879]: Connection closed by 10.0.0.1 port 52640 Jan 21 05:53:52.709477 sshd-session[1875]: pam_unix(sshd:session): session closed for user core Jan 21 05:53:52.719000 audit[1875]: USER_END pid=1875 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:53:52.736858 systemd[1]: sshd@8-10.0.0.61:22-10.0.0.1:52640.service: Deactivated successfully. Jan 21 05:53:52.719000 audit[1875]: CRED_DISP pid=1875 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:53:52.745247 systemd[1]: session-10.scope: Deactivated successfully. Jan 21 05:53:52.746063 systemd[1]: session-10.scope: Consumed 17.372s CPU time, 216.6M memory peak. Jan 21 05:53:52.754209 systemd-logind[1620]: Session 10 logged out. Waiting for processes to exit. Jan 21 05:53:52.756337 kernel: audit: type=1106 audit(1768974832.719:520): pid=1875 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:53:52.756987 kernel: audit: type=1104 audit(1768974832.719:521): pid=1875 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:53:52.734000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.61:22-10.0.0.1:52640 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:53:52.757885 systemd-logind[1620]: Removed session 10. Jan 21 05:53:52.771812 kernel: audit: type=1131 audit(1768974832.734:522): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.61:22-10.0.0.1:52640 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:54:00.653000 audit[3404]: NETFILTER_CFG table=filter:105 family=2 entries=15 op=nft_register_rule pid=3404 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:00.665464 kernel: audit: type=1325 audit(1768974840.653:523): table=filter:105 family=2 entries=15 op=nft_register_rule pid=3404 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:00.688787 kernel: audit: type=1300 audit(1768974840.653:523): arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffcc69204c0 a2=0 a3=7ffcc69204ac items=0 ppid=3144 pid=3404 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:00.653000 audit[3404]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffcc69204c0 a2=0 a3=7ffcc69204ac items=0 ppid=3144 pid=3404 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:00.653000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:00.717783 kernel: audit: type=1327 audit(1768974840.653:523): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:00.717936 kernel: audit: type=1325 audit(1768974840.664:524): table=nat:106 family=2 entries=12 op=nft_register_rule pid=3404 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:00.717979 kernel: audit: type=1300 audit(1768974840.664:524): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffcc69204c0 a2=0 a3=0 items=0 ppid=3144 pid=3404 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:00.664000 audit[3404]: NETFILTER_CFG table=nat:106 family=2 entries=12 op=nft_register_rule pid=3404 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:00.664000 audit[3404]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffcc69204c0 a2=0 a3=0 items=0 ppid=3144 pid=3404 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:00.742240 kernel: audit: type=1327 audit(1768974840.664:524): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:00.664000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:00.757241 kernel: audit: type=1325 audit(1768974840.741:525): table=filter:107 family=2 entries=16 op=nft_register_rule pid=3406 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:00.741000 audit[3406]: NETFILTER_CFG table=filter:107 family=2 entries=16 op=nft_register_rule pid=3406 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:00.741000 audit[3406]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffc70b57090 a2=0 a3=7ffc70b5707c items=0 ppid=3144 pid=3406 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:00.796553 kernel: audit: type=1300 audit(1768974840.741:525): arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffc70b57090 a2=0 a3=7ffc70b5707c items=0 ppid=3144 pid=3406 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:00.797596 kernel: audit: type=1327 audit(1768974840.741:525): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:00.741000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:00.817244 kernel: audit: type=1325 audit(1768974840.799:526): table=nat:108 family=2 entries=12 op=nft_register_rule pid=3406 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:00.799000 audit[3406]: NETFILTER_CFG table=nat:108 family=2 entries=12 op=nft_register_rule pid=3406 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:00.799000 audit[3406]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc70b57090 a2=0 a3=0 items=0 ppid=3144 pid=3406 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:00.799000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:22.170115 kernel: kauditd_printk_skb: 2 callbacks suppressed Jan 21 05:54:22.180355 kernel: audit: type=1325 audit(1768974862.144:527): table=filter:109 family=2 entries=17 op=nft_register_rule pid=3410 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:22.144000 audit[3410]: NETFILTER_CFG table=filter:109 family=2 entries=17 op=nft_register_rule pid=3410 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:22.144000 audit[3410]: SYSCALL arch=c000003e syscall=46 success=yes exit=6736 a0=3 a1=7ffdca7a0050 a2=0 a3=7ffdca7a003c items=0 ppid=3144 pid=3410 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:22.288511 kernel: audit: type=1300 audit(1768974862.144:527): arch=c000003e syscall=46 success=yes exit=6736 a0=3 a1=7ffdca7a0050 a2=0 a3=7ffdca7a003c items=0 ppid=3144 pid=3410 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:22.289439 kernel: audit: type=1327 audit(1768974862.144:527): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:22.144000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:22.323791 kernel: audit: type=1325 audit(1768974862.191:528): table=nat:110 family=2 entries=12 op=nft_register_rule pid=3410 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:22.191000 audit[3410]: NETFILTER_CFG table=nat:110 family=2 entries=12 op=nft_register_rule pid=3410 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:22.351824 kernel: audit: type=1300 audit(1768974862.191:528): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffdca7a0050 a2=0 a3=0 items=0 ppid=3144 pid=3410 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:22.191000 audit[3410]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffdca7a0050 a2=0 a3=0 items=0 ppid=3144 pid=3410 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:22.191000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:22.459743 kernel: audit: type=1327 audit(1768974862.191:528): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:22.657000 audit[3412]: NETFILTER_CFG table=filter:111 family=2 entries=19 op=nft_register_rule pid=3412 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:22.693356 kernel: audit: type=1325 audit(1768974862.657:529): table=filter:111 family=2 entries=19 op=nft_register_rule pid=3412 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:22.657000 audit[3412]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffd90c6d100 a2=0 a3=7ffd90c6d0ec items=0 ppid=3144 pid=3412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:22.750173 kernel: audit: type=1300 audit(1768974862.657:529): arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffd90c6d100 a2=0 a3=7ffd90c6d0ec items=0 ppid=3144 pid=3412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:22.657000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:22.843225 kernel: audit: type=1327 audit(1768974862.657:529): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:22.769000 audit[3412]: NETFILTER_CFG table=nat:112 family=2 entries=12 op=nft_register_rule pid=3412 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:22.891216 kernel: audit: type=1325 audit(1768974862.769:530): table=nat:112 family=2 entries=12 op=nft_register_rule pid=3412 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:22.769000 audit[3412]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd90c6d100 a2=0 a3=0 items=0 ppid=3144 pid=3412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:22.769000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:31.510000 audit[3420]: NETFILTER_CFG table=filter:113 family=2 entries=21 op=nft_register_rule pid=3420 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:31.523880 kernel: kauditd_printk_skb: 2 callbacks suppressed Jan 21 05:54:31.524005 kernel: audit: type=1325 audit(1768974871.510:531): table=filter:113 family=2 entries=21 op=nft_register_rule pid=3420 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:31.592087 kernel: audit: type=1300 audit(1768974871.510:531): arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffd8fb91e10 a2=0 a3=7ffd8fb91dfc items=0 ppid=3144 pid=3420 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:31.510000 audit[3420]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffd8fb91e10 a2=0 a3=7ffd8fb91dfc items=0 ppid=3144 pid=3420 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:31.510000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:31.756110 kernel: audit: type=1327 audit(1768974871.510:531): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:31.756342 kernel: audit: type=1325 audit(1768974871.591:532): table=nat:114 family=2 entries=12 op=nft_register_rule pid=3420 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:31.591000 audit[3420]: NETFILTER_CFG table=nat:114 family=2 entries=12 op=nft_register_rule pid=3420 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:31.591000 audit[3420]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd8fb91e10 a2=0 a3=0 items=0 ppid=3144 pid=3420 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:31.846210 kernel: audit: type=1300 audit(1768974871.591:532): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd8fb91e10 a2=0 a3=0 items=0 ppid=3144 pid=3420 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:31.883573 kernel: audit: type=1327 audit(1768974871.591:532): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:31.591000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:31.984000 audit[3422]: NETFILTER_CFG table=filter:115 family=2 entries=22 op=nft_register_rule pid=3422 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:32.056084 kernel: audit: type=1325 audit(1768974871.984:533): table=filter:115 family=2 entries=22 op=nft_register_rule pid=3422 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:32.094623 kernel: audit: type=1300 audit(1768974871.984:533): arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffc09c8e800 a2=0 a3=7ffc09c8e7ec items=0 ppid=3144 pid=3422 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:32.110559 kernel: audit: type=1327 audit(1768974871.984:533): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:31.984000 audit[3422]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffc09c8e800 a2=0 a3=7ffc09c8e7ec items=0 ppid=3144 pid=3422 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:31.984000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:32.128000 audit[3422]: NETFILTER_CFG table=nat:116 family=2 entries=12 op=nft_register_rule pid=3422 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:32.190563 kernel: audit: type=1325 audit(1768974872.128:534): table=nat:116 family=2 entries=12 op=nft_register_rule pid=3422 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:32.128000 audit[3422]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc09c8e800 a2=0 a3=0 items=0 ppid=3144 pid=3422 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:32.128000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:32.232201 kubelet[2982]: I0121 05:54:32.229418 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82c5605c-5598-4ab0-8d52-e8fdaaac2be0-tigera-ca-bundle\") pod \"calico-typha-5596648d99-m7djd\" (UID: \"82c5605c-5598-4ab0-8d52-e8fdaaac2be0\") " pod="calico-system/calico-typha-5596648d99-m7djd" Jan 21 05:54:32.232201 kubelet[2982]: I0121 05:54:32.229552 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjkbq\" (UniqueName: \"kubernetes.io/projected/82c5605c-5598-4ab0-8d52-e8fdaaac2be0-kube-api-access-hjkbq\") pod \"calico-typha-5596648d99-m7djd\" (UID: \"82c5605c-5598-4ab0-8d52-e8fdaaac2be0\") " pod="calico-system/calico-typha-5596648d99-m7djd" Jan 21 05:54:32.232201 kubelet[2982]: I0121 05:54:32.229582 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/82c5605c-5598-4ab0-8d52-e8fdaaac2be0-typha-certs\") pod \"calico-typha-5596648d99-m7djd\" (UID: \"82c5605c-5598-4ab0-8d52-e8fdaaac2be0\") " pod="calico-system/calico-typha-5596648d99-m7djd" Jan 21 05:54:32.382746 systemd[1]: Created slice kubepods-besteffort-pod82c5605c_5598_4ab0_8d52_e8fdaaac2be0.slice - libcontainer container kubepods-besteffort-pod82c5605c_5598_4ab0_8d52_e8fdaaac2be0.slice. Jan 21 05:54:32.705000 audit[3426]: NETFILTER_CFG table=filter:117 family=2 entries=22 op=nft_register_rule pid=3426 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:32.705000 audit[3426]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7fffe5bfb900 a2=0 a3=7fffe5bfb8ec items=0 ppid=3144 pid=3426 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:32.705000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:32.721000 audit[3426]: NETFILTER_CFG table=nat:118 family=2 entries=12 op=nft_register_rule pid=3426 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:32.721000 audit[3426]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fffe5bfb900 a2=0 a3=0 items=0 ppid=3144 pid=3426 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:32.721000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:33.088420 kubelet[2982]: E0121 05:54:33.088275 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:33.108429 containerd[1639]: time="2026-01-21T05:54:33.108269259Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5596648d99-m7djd,Uid:82c5605c-5598-4ab0-8d52-e8fdaaac2be0,Namespace:calico-system,Attempt:0,}" Jan 21 05:54:33.449407 kubelet[2982]: I0121 05:54:33.448258 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-cni-bin-dir\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.449407 kubelet[2982]: I0121 05:54:33.448320 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-cni-net-dir\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.449407 kubelet[2982]: I0121 05:54:33.448344 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-var-run-calico\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.449407 kubelet[2982]: I0121 05:54:33.448367 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-policysync\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.449407 kubelet[2982]: I0121 05:54:33.448390 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-lib-modules\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.454549 kubelet[2982]: I0121 05:54:33.448414 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-cni-log-dir\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.454549 kubelet[2982]: I0121 05:54:33.448441 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-flexvol-driver-host\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.454549 kubelet[2982]: I0121 05:54:33.448468 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-tigera-ca-bundle\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.454549 kubelet[2982]: I0121 05:54:33.448492 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-node-certs\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.454549 kubelet[2982]: I0121 05:54:33.448513 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-var-lib-calico\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.465150 kubelet[2982]: I0121 05:54:33.448538 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-xtables-lock\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.465150 kubelet[2982]: I0121 05:54:33.448564 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4jjh\" (UniqueName: \"kubernetes.io/projected/e244fc2a-8aed-4ba9-8211-4cf0914c46b1-kube-api-access-s4jjh\") pod \"calico-node-tscdz\" (UID: \"e244fc2a-8aed-4ba9-8211-4cf0914c46b1\") " pod="calico-system/calico-node-tscdz" Jan 21 05:54:33.478611 systemd[1]: Created slice kubepods-besteffort-pode244fc2a_8aed_4ba9_8211_4cf0914c46b1.slice - libcontainer container kubepods-besteffort-pode244fc2a_8aed_4ba9_8211_4cf0914c46b1.slice. Jan 21 05:54:33.538343 kubelet[2982]: E0121 05:54:33.537807 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:33.584239 kubelet[2982]: E0121 05:54:33.583868 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.584239 kubelet[2982]: W0121 05:54:33.583972 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.612477 containerd[1639]: time="2026-01-21T05:54:33.610224799Z" level=info msg="connecting to shim a812ac7a46cdd6272a729c7bf8355e2864a58adb68f334d1825155c4e53dd22e" address="unix:///run/containerd/s/d790881195b101fa9e356752fea53ebfb66421b00a10d1ccc5a1707fcc70c91f" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:54:33.687810 kubelet[2982]: E0121 05:54:33.686943 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.742460 kubelet[2982]: E0121 05:54:33.733054 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.742460 kubelet[2982]: W0121 05:54:33.742300 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.742460 kubelet[2982]: E0121 05:54:33.742340 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.774569 kubelet[2982]: E0121 05:54:33.772429 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.774569 kubelet[2982]: W0121 05:54:33.772536 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.774569 kubelet[2982]: E0121 05:54:33.772571 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.778974 kubelet[2982]: E0121 05:54:33.778866 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.779436 kubelet[2982]: W0121 05:54:33.778974 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.779436 kubelet[2982]: E0121 05:54:33.779006 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.780472 kubelet[2982]: E0121 05:54:33.780374 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.780472 kubelet[2982]: W0121 05:54:33.780469 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.780574 kubelet[2982]: E0121 05:54:33.780493 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.796959 kubelet[2982]: E0121 05:54:33.789325 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.796959 kubelet[2982]: W0121 05:54:33.789431 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.796959 kubelet[2982]: E0121 05:54:33.789463 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.839923 kubelet[2982]: E0121 05:54:33.839610 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.839923 kubelet[2982]: W0121 05:54:33.839815 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.847913 kubelet[2982]: E0121 05:54:33.842259 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.847913 kubelet[2982]: E0121 05:54:33.842865 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.847913 kubelet[2982]: W0121 05:54:33.842880 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.847913 kubelet[2982]: E0121 05:54:33.842897 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.847913 kubelet[2982]: I0121 05:54:33.844159 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/487c6aa2-04b5-4684-9420-0e4b73799fd3-registration-dir\") pod \"csi-node-driver-zftdp\" (UID: \"487c6aa2-04b5-4684-9420-0e4b73799fd3\") " pod="calico-system/csi-node-driver-zftdp" Jan 21 05:54:33.847913 kubelet[2982]: E0121 05:54:33.846785 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.847913 kubelet[2982]: W0121 05:54:33.846808 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.847913 kubelet[2982]: E0121 05:54:33.846831 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.887000 kubelet[2982]: E0121 05:54:33.851962 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.887000 kubelet[2982]: W0121 05:54:33.851982 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.887000 kubelet[2982]: E0121 05:54:33.852001 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.887000 kubelet[2982]: E0121 05:54:33.852530 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.887000 kubelet[2982]: W0121 05:54:33.852544 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.887000 kubelet[2982]: E0121 05:54:33.852558 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.887000 kubelet[2982]: E0121 05:54:33.856789 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.887000 kubelet[2982]: W0121 05:54:33.856815 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.887000 kubelet[2982]: E0121 05:54:33.856834 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.887000 kubelet[2982]: E0121 05:54:33.860999 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.937476 kubelet[2982]: W0121 05:54:33.861018 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.937476 kubelet[2982]: E0121 05:54:33.861036 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.937476 kubelet[2982]: E0121 05:54:33.865619 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.937476 kubelet[2982]: W0121 05:54:33.866440 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.937476 kubelet[2982]: E0121 05:54:33.866466 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.937476 kubelet[2982]: E0121 05:54:33.866980 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.937476 kubelet[2982]: W0121 05:54:33.866992 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.937476 kubelet[2982]: E0121 05:54:33.867012 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.937476 kubelet[2982]: E0121 05:54:33.867438 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.937476 kubelet[2982]: W0121 05:54:33.867451 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.939417 kubelet[2982]: E0121 05:54:33.867464 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.939417 kubelet[2982]: I0121 05:54:33.873247 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/487c6aa2-04b5-4684-9420-0e4b73799fd3-socket-dir\") pod \"csi-node-driver-zftdp\" (UID: \"487c6aa2-04b5-4684-9420-0e4b73799fd3\") " pod="calico-system/csi-node-driver-zftdp" Jan 21 05:54:33.939417 kubelet[2982]: E0121 05:54:33.874469 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.939417 kubelet[2982]: W0121 05:54:33.874497 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.939417 kubelet[2982]: E0121 05:54:33.874526 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.939417 kubelet[2982]: E0121 05:54:33.875057 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.939417 kubelet[2982]: W0121 05:54:33.880342 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.939417 kubelet[2982]: E0121 05:54:33.880389 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.939417 kubelet[2982]: E0121 05:54:33.881359 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.941392 kubelet[2982]: W0121 05:54:33.881375 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.941392 kubelet[2982]: E0121 05:54:33.881390 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.941392 kubelet[2982]: E0121 05:54:33.883021 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.941392 kubelet[2982]: W0121 05:54:33.883035 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.941392 kubelet[2982]: E0121 05:54:33.883050 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.941392 kubelet[2982]: E0121 05:54:33.886575 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.941392 kubelet[2982]: W0121 05:54:33.886966 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.941392 kubelet[2982]: E0121 05:54:33.886988 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.941392 kubelet[2982]: E0121 05:54:33.891870 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.941392 kubelet[2982]: W0121 05:54:33.891899 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.942561 kubelet[2982]: E0121 05:54:33.891939 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.942561 kubelet[2982]: I0121 05:54:33.893449 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgtqr\" (UniqueName: \"kubernetes.io/projected/487c6aa2-04b5-4684-9420-0e4b73799fd3-kube-api-access-qgtqr\") pod \"csi-node-driver-zftdp\" (UID: \"487c6aa2-04b5-4684-9420-0e4b73799fd3\") " pod="calico-system/csi-node-driver-zftdp" Jan 21 05:54:33.942561 kubelet[2982]: E0121 05:54:33.901471 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.942561 kubelet[2982]: W0121 05:54:33.901491 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.942561 kubelet[2982]: E0121 05:54:33.901511 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.942561 kubelet[2982]: E0121 05:54:33.902877 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.942561 kubelet[2982]: W0121 05:54:33.902896 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.942561 kubelet[2982]: E0121 05:54:33.902916 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.942561 kubelet[2982]: E0121 05:54:33.918515 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.947170 kubelet[2982]: W0121 05:54:33.918533 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.947170 kubelet[2982]: E0121 05:54:33.918562 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.947170 kubelet[2982]: E0121 05:54:33.920182 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.947170 kubelet[2982]: W0121 05:54:33.920199 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.947170 kubelet[2982]: E0121 05:54:33.920218 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.947170 kubelet[2982]: E0121 05:54:33.929185 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.947170 kubelet[2982]: W0121 05:54:33.929210 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.947170 kubelet[2982]: E0121 05:54:33.929235 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.947170 kubelet[2982]: E0121 05:54:33.930032 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.947170 kubelet[2982]: W0121 05:54:33.930046 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.947509 kubelet[2982]: E0121 05:54:33.930060 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.947509 kubelet[2982]: E0121 05:54:33.941186 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.947509 kubelet[2982]: W0121 05:54:33.941208 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.947509 kubelet[2982]: E0121 05:54:33.941233 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.947509 kubelet[2982]: I0121 05:54:33.947434 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/487c6aa2-04b5-4684-9420-0e4b73799fd3-kubelet-dir\") pod \"csi-node-driver-zftdp\" (UID: \"487c6aa2-04b5-4684-9420-0e4b73799fd3\") " pod="calico-system/csi-node-driver-zftdp" Jan 21 05:54:33.967581 kubelet[2982]: E0121 05:54:33.961350 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.967581 kubelet[2982]: W0121 05:54:33.961389 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.967581 kubelet[2982]: E0121 05:54:33.961421 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.968966 kubelet[2982]: E0121 05:54:33.968814 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.968966 kubelet[2982]: W0121 05:54:33.968912 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.968966 kubelet[2982]: E0121 05:54:33.968941 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:33.978173 kubelet[2982]: E0121 05:54:33.973464 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:33.978173 kubelet[2982]: W0121 05:54:33.973801 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:33.978173 kubelet[2982]: E0121 05:54:33.973828 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.008492 kubelet[2982]: E0121 05:54:34.005212 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.008492 kubelet[2982]: W0121 05:54:34.005245 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.008492 kubelet[2982]: E0121 05:54:34.005277 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.008492 kubelet[2982]: I0121 05:54:34.005326 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/487c6aa2-04b5-4684-9420-0e4b73799fd3-varrun\") pod \"csi-node-driver-zftdp\" (UID: \"487c6aa2-04b5-4684-9420-0e4b73799fd3\") " pod="calico-system/csi-node-driver-zftdp" Jan 21 05:54:34.035929 kubelet[2982]: E0121 05:54:34.030445 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.035929 kubelet[2982]: W0121 05:54:34.030586 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.035929 kubelet[2982]: E0121 05:54:34.030621 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.059998 kubelet[2982]: E0121 05:54:34.059549 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.059998 kubelet[2982]: W0121 05:54:34.059839 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.059998 kubelet[2982]: E0121 05:54:34.059880 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.096872 kubelet[2982]: E0121 05:54:34.095435 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.096872 kubelet[2982]: W0121 05:54:34.095534 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.096872 kubelet[2982]: E0121 05:54:34.095568 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.125953 kubelet[2982]: E0121 05:54:34.121919 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.125953 kubelet[2982]: W0121 05:54:34.121965 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.125953 kubelet[2982]: E0121 05:54:34.122000 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.159839 kubelet[2982]: E0121 05:54:34.159298 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.159839 kubelet[2982]: W0121 05:54:34.159600 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.159839 kubelet[2982]: E0121 05:54:34.159791 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.172784 kubelet[2982]: E0121 05:54:34.172398 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.173008 kubelet[2982]: W0121 05:54:34.172982 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.173779 kubelet[2982]: E0121 05:54:34.173101 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.181781 kubelet[2982]: E0121 05:54:34.181359 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.184828 kubelet[2982]: W0121 05:54:34.184483 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.184828 kubelet[2982]: E0121 05:54:34.184533 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.192866 kubelet[2982]: E0121 05:54:34.192828 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.202278 kubelet[2982]: W0121 05:54:34.202220 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.202543 kubelet[2982]: E0121 05:54:34.202514 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.295200 kubelet[2982]: E0121 05:54:34.261077 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.295200 kubelet[2982]: W0121 05:54:34.276562 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.295200 kubelet[2982]: E0121 05:54:34.277840 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.295200 kubelet[2982]: E0121 05:54:34.290277 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.295200 kubelet[2982]: W0121 05:54:34.290310 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.295200 kubelet[2982]: E0121 05:54:34.290343 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.295885 kubelet[2982]: E0121 05:54:34.295456 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.295885 kubelet[2982]: W0121 05:54:34.295482 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.295885 kubelet[2982]: E0121 05:54:34.295511 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.303216 kubelet[2982]: E0121 05:54:34.302121 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.303216 kubelet[2982]: W0121 05:54:34.302236 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.303216 kubelet[2982]: E0121 05:54:34.302266 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.306608 kubelet[2982]: E0121 05:54:34.306259 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.306608 kubelet[2982]: W0121 05:54:34.306280 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.306608 kubelet[2982]: E0121 05:54:34.306307 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.314625 kubelet[2982]: E0121 05:54:34.307618 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.314625 kubelet[2982]: W0121 05:54:34.307633 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.314625 kubelet[2982]: E0121 05:54:34.307801 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.314625 kubelet[2982]: E0121 05:54:34.308451 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.314625 kubelet[2982]: W0121 05:54:34.308464 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.314625 kubelet[2982]: E0121 05:54:34.308477 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.314625 kubelet[2982]: E0121 05:54:34.308901 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.314625 kubelet[2982]: W0121 05:54:34.308912 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.314625 kubelet[2982]: E0121 05:54:34.308924 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.314625 kubelet[2982]: E0121 05:54:34.309272 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.318978 kubelet[2982]: W0121 05:54:34.309286 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.318978 kubelet[2982]: E0121 05:54:34.309298 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.318978 kubelet[2982]: E0121 05:54:34.309568 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.318978 kubelet[2982]: W0121 05:54:34.309580 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.318978 kubelet[2982]: E0121 05:54:34.309593 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.318978 kubelet[2982]: E0121 05:54:34.310064 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.318978 kubelet[2982]: W0121 05:54:34.310078 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.318978 kubelet[2982]: E0121 05:54:34.310092 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.337804 kubelet[2982]: E0121 05:54:34.336369 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.337804 kubelet[2982]: W0121 05:54:34.336414 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.337804 kubelet[2982]: E0121 05:54:34.336457 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.340083 kubelet[2982]: E0121 05:54:34.339956 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.340083 kubelet[2982]: W0121 05:54:34.340060 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.340306 kubelet[2982]: E0121 05:54:34.340096 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.368854 kubelet[2982]: E0121 05:54:34.365061 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.368854 kubelet[2982]: W0121 05:54:34.365099 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.368854 kubelet[2982]: E0121 05:54:34.365218 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.386958 kubelet[2982]: E0121 05:54:34.386234 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.386958 kubelet[2982]: W0121 05:54:34.386559 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.386958 kubelet[2982]: E0121 05:54:34.386598 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.397542 kubelet[2982]: E0121 05:54:34.394887 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.397542 kubelet[2982]: W0121 05:54:34.394922 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.397542 kubelet[2982]: E0121 05:54:34.394953 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.397542 kubelet[2982]: E0121 05:54:34.396871 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.397542 kubelet[2982]: W0121 05:54:34.396887 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.397542 kubelet[2982]: E0121 05:54:34.396904 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.408075 kubelet[2982]: E0121 05:54:34.406985 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.408075 kubelet[2982]: W0121 05:54:34.407022 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.408075 kubelet[2982]: E0121 05:54:34.407052 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.412127 kubelet[2982]: E0121 05:54:34.411471 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.412127 kubelet[2982]: W0121 05:54:34.411496 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.412127 kubelet[2982]: E0121 05:54:34.411521 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.416263 kubelet[2982]: E0121 05:54:34.415260 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.420083 kubelet[2982]: W0121 05:54:34.420049 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.420861 kubelet[2982]: E0121 05:54:34.420288 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.423088 kubelet[2982]: E0121 05:54:34.423000 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:34.436882 kubelet[2982]: E0121 05:54:34.434533 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.436882 kubelet[2982]: W0121 05:54:34.434573 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.436882 kubelet[2982]: E0121 05:54:34.434608 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.437132 containerd[1639]: time="2026-01-21T05:54:34.435920797Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-tscdz,Uid:e244fc2a-8aed-4ba9-8211-4cf0914c46b1,Namespace:calico-system,Attempt:0,}" Jan 21 05:54:34.442316 kubelet[2982]: E0121 05:54:34.442137 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.442316 kubelet[2982]: W0121 05:54:34.442239 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.442316 kubelet[2982]: E0121 05:54:34.442261 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.651554 kubelet[2982]: E0121 05:54:34.644067 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:34.653458 kubelet[2982]: W0121 05:54:34.652594 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:34.653914 kubelet[2982]: E0121 05:54:34.653324 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:34.711455 systemd[1]: Started cri-containerd-a812ac7a46cdd6272a729c7bf8355e2864a58adb68f334d1825155c4e53dd22e.scope - libcontainer container a812ac7a46cdd6272a729c7bf8355e2864a58adb68f334d1825155c4e53dd22e. Jan 21 05:54:34.775021 containerd[1639]: time="2026-01-21T05:54:34.774955862Z" level=info msg="connecting to shim e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c" address="unix:///run/containerd/s/a097f43dd299e2930c4a70b814880e743e86b92e44c5686bf5acf59179f1242e" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:54:34.966000 audit: BPF prog-id=149 op=LOAD Jan 21 05:54:34.972000 audit: BPF prog-id=150 op=LOAD Jan 21 05:54:34.972000 audit[3492]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c238 a2=98 a3=0 items=0 ppid=3436 pid=3492 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:34.972000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6138313261633761343663646436323732613732396337626638333535 Jan 21 05:54:34.972000 audit: BPF prog-id=150 op=UNLOAD Jan 21 05:54:34.972000 audit[3492]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3436 pid=3492 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:34.972000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6138313261633761343663646436323732613732396337626638333535 Jan 21 05:54:34.981000 audit: BPF prog-id=151 op=LOAD Jan 21 05:54:34.981000 audit[3492]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c488 a2=98 a3=0 items=0 ppid=3436 pid=3492 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:34.981000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6138313261633761343663646436323732613732396337626638333535 Jan 21 05:54:35.008000 audit: BPF prog-id=152 op=LOAD Jan 21 05:54:35.008000 audit[3492]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00010c218 a2=98 a3=0 items=0 ppid=3436 pid=3492 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.008000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6138313261633761343663646436323732613732396337626638333535 Jan 21 05:54:35.047000 audit: BPF prog-id=152 op=UNLOAD Jan 21 05:54:35.047000 audit[3492]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3436 pid=3492 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.047000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6138313261633761343663646436323732613732396337626638333535 Jan 21 05:54:35.058000 audit: BPF prog-id=151 op=UNLOAD Jan 21 05:54:35.058000 audit[3492]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3436 pid=3492 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.058000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6138313261633761343663646436323732613732396337626638333535 Jan 21 05:54:35.071000 audit: BPF prog-id=153 op=LOAD Jan 21 05:54:35.071000 audit[3492]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c6e8 a2=98 a3=0 items=0 ppid=3436 pid=3492 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.071000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6138313261633761343663646436323732613732396337626638333535 Jan 21 05:54:35.214396 kubelet[2982]: E0121 05:54:35.210231 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:35.354787 systemd[1]: Started cri-containerd-e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c.scope - libcontainer container e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c. Jan 21 05:54:35.508000 audit: BPF prog-id=154 op=LOAD Jan 21 05:54:35.515000 audit: BPF prog-id=155 op=LOAD Jan 21 05:54:35.515000 audit[3560]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=3542 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.515000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6530363736326438373264313138616266386633393832333533623066 Jan 21 05:54:35.515000 audit: BPF prog-id=155 op=UNLOAD Jan 21 05:54:35.515000 audit[3560]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.515000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6530363736326438373264313138616266386633393832333533623066 Jan 21 05:54:35.516000 audit: BPF prog-id=156 op=LOAD Jan 21 05:54:35.516000 audit[3560]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3542 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.516000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6530363736326438373264313138616266386633393832333533623066 Jan 21 05:54:35.516000 audit: BPF prog-id=157 op=LOAD Jan 21 05:54:35.516000 audit[3560]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=3542 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.516000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6530363736326438373264313138616266386633393832333533623066 Jan 21 05:54:35.516000 audit: BPF prog-id=157 op=UNLOAD Jan 21 05:54:35.516000 audit[3560]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.516000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6530363736326438373264313138616266386633393832333533623066 Jan 21 05:54:35.516000 audit: BPF prog-id=156 op=UNLOAD Jan 21 05:54:35.516000 audit[3560]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.516000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6530363736326438373264313138616266386633393832333533623066 Jan 21 05:54:35.517000 audit: BPF prog-id=158 op=LOAD Jan 21 05:54:35.517000 audit[3560]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=3542 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:35.517000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6530363736326438373264313138616266386633393832333533623066 Jan 21 05:54:35.717594 containerd[1639]: time="2026-01-21T05:54:35.711525358Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5596648d99-m7djd,Uid:82c5605c-5598-4ab0-8d52-e8fdaaac2be0,Namespace:calico-system,Attempt:0,} returns sandbox id \"a812ac7a46cdd6272a729c7bf8355e2864a58adb68f334d1825155c4e53dd22e\"" Jan 21 05:54:35.743139 kubelet[2982]: E0121 05:54:35.721287 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:35.753501 containerd[1639]: time="2026-01-21T05:54:35.728079205Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Jan 21 05:54:35.985175 containerd[1639]: time="2026-01-21T05:54:35.985101083Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-tscdz,Uid:e244fc2a-8aed-4ba9-8211-4cf0914c46b1,Namespace:calico-system,Attempt:0,} returns sandbox id \"e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c\"" Jan 21 05:54:35.990454 kubelet[2982]: E0121 05:54:35.986937 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:36.263972 kubelet[2982]: E0121 05:54:36.260004 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:36.326281 kubelet[2982]: E0121 05:54:36.325967 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.326281 kubelet[2982]: W0121 05:54:36.326002 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.326281 kubelet[2982]: E0121 05:54:36.326030 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.328844 kubelet[2982]: E0121 05:54:36.328543 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.328844 kubelet[2982]: W0121 05:54:36.328561 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.328844 kubelet[2982]: E0121 05:54:36.328585 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.400364 kubelet[2982]: E0121 05:54:36.395303 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.427910 kubelet[2982]: W0121 05:54:36.402269 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.427910 kubelet[2982]: E0121 05:54:36.402312 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.427910 kubelet[2982]: E0121 05:54:36.412798 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.427910 kubelet[2982]: W0121 05:54:36.412822 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.427910 kubelet[2982]: E0121 05:54:36.412850 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.427910 kubelet[2982]: E0121 05:54:36.424315 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.427910 kubelet[2982]: W0121 05:54:36.424346 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.427910 kubelet[2982]: E0121 05:54:36.424455 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.428553 kubelet[2982]: E0121 05:54:36.428086 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.428553 kubelet[2982]: W0121 05:54:36.428103 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.428553 kubelet[2982]: E0121 05:54:36.428120 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.431264 kubelet[2982]: E0121 05:54:36.429261 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.431264 kubelet[2982]: W0121 05:54:36.429273 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.431264 kubelet[2982]: E0121 05:54:36.429286 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.431264 kubelet[2982]: E0121 05:54:36.430486 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.431264 kubelet[2982]: W0121 05:54:36.430501 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.431264 kubelet[2982]: E0121 05:54:36.430515 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.440768 kubelet[2982]: E0121 05:54:36.440515 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.440768 kubelet[2982]: W0121 05:54:36.440543 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.440768 kubelet[2982]: E0121 05:54:36.440567 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.448922 kubelet[2982]: E0121 05:54:36.446628 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.448922 kubelet[2982]: W0121 05:54:36.446782 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.448922 kubelet[2982]: E0121 05:54:36.446803 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.448922 kubelet[2982]: E0121 05:54:36.448139 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.448922 kubelet[2982]: W0121 05:54:36.448152 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.448922 kubelet[2982]: E0121 05:54:36.448165 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.451831 kubelet[2982]: E0121 05:54:36.449955 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.451831 kubelet[2982]: W0121 05:54:36.450027 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.451831 kubelet[2982]: E0121 05:54:36.450043 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.454105 kubelet[2982]: E0121 05:54:36.454002 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.454105 kubelet[2982]: W0121 05:54:36.454023 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.454105 kubelet[2982]: E0121 05:54:36.454038 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.457255 kubelet[2982]: E0121 05:54:36.457148 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.457255 kubelet[2982]: W0121 05:54:36.457171 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.457255 kubelet[2982]: E0121 05:54:36.457187 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.463023 kubelet[2982]: E0121 05:54:36.462785 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.463023 kubelet[2982]: W0121 05:54:36.462805 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.463023 kubelet[2982]: E0121 05:54:36.462823 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.465530 kubelet[2982]: E0121 05:54:36.464557 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.465530 kubelet[2982]: W0121 05:54:36.464573 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.465530 kubelet[2982]: E0121 05:54:36.464592 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.467372 kubelet[2982]: E0121 05:54:36.467118 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.467372 kubelet[2982]: W0121 05:54:36.467131 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.467372 kubelet[2982]: E0121 05:54:36.467144 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.471168 kubelet[2982]: E0121 05:54:36.470995 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.471168 kubelet[2982]: W0121 05:54:36.471069 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.471168 kubelet[2982]: E0121 05:54:36.471088 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.476930 kubelet[2982]: E0121 05:54:36.474837 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.476930 kubelet[2982]: W0121 05:54:36.474918 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.476930 kubelet[2982]: E0121 05:54:36.474936 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.481760 kubelet[2982]: E0121 05:54:36.480969 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.481760 kubelet[2982]: W0121 05:54:36.481053 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.481760 kubelet[2982]: E0121 05:54:36.481074 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.486254 kubelet[2982]: E0121 05:54:36.485856 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.486254 kubelet[2982]: W0121 05:54:36.485944 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.486254 kubelet[2982]: E0121 05:54:36.485963 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.494298 kubelet[2982]: E0121 05:54:36.494139 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.494298 kubelet[2982]: W0121 05:54:36.494168 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.494298 kubelet[2982]: E0121 05:54:36.494194 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.500477 kubelet[2982]: E0121 05:54:36.499994 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.500477 kubelet[2982]: W0121 05:54:36.500078 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.500477 kubelet[2982]: E0121 05:54:36.500102 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.505764 kubelet[2982]: E0121 05:54:36.505056 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.505764 kubelet[2982]: W0121 05:54:36.505137 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.505764 kubelet[2982]: E0121 05:54:36.505161 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.507916 kubelet[2982]: E0121 05:54:36.506368 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.507916 kubelet[2982]: W0121 05:54:36.506384 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.507916 kubelet[2982]: E0121 05:54:36.506480 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.509373 kubelet[2982]: E0121 05:54:36.509275 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.509373 kubelet[2982]: W0121 05:54:36.509292 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.509373 kubelet[2982]: E0121 05:54:36.509308 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.513376 kubelet[2982]: E0121 05:54:36.513354 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.513783 kubelet[2982]: W0121 05:54:36.513564 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.513783 kubelet[2982]: E0121 05:54:36.513590 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.521846 kubelet[2982]: E0121 05:54:36.518190 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.521846 kubelet[2982]: W0121 05:54:36.518211 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.521846 kubelet[2982]: E0121 05:54:36.518228 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.521846 kubelet[2982]: E0121 05:54:36.520591 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.522525 kubelet[2982]: W0121 05:54:36.520605 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.522525 kubelet[2982]: E0121 05:54:36.522502 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.539087 kubelet[2982]: E0121 05:54:36.539047 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.540863 kubelet[2982]: W0121 05:54:36.539233 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.540863 kubelet[2982]: E0121 05:54:36.539268 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.545074 kubelet[2982]: E0121 05:54:36.544535 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.545074 kubelet[2982]: W0121 05:54:36.544559 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.545074 kubelet[2982]: E0121 05:54:36.544581 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:36.553201 kubelet[2982]: E0121 05:54:36.552102 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:36.553201 kubelet[2982]: W0121 05:54:36.552120 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:36.553201 kubelet[2982]: E0121 05:54:36.552142 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:37.217444 kubelet[2982]: E0121 05:54:37.217381 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:39.332170 kubelet[2982]: E0121 05:54:39.331589 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:39.364586 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount114776812.mount: Deactivated successfully. Jan 21 05:54:41.231863 kubelet[2982]: E0121 05:54:41.210257 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:42.216551 kubelet[2982]: E0121 05:54:42.215906 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:42.228397 kubelet[2982]: E0121 05:54:42.228356 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.228397 kubelet[2982]: W0121 05:54:42.228389 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.228600 kubelet[2982]: E0121 05:54:42.228420 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.234612 kubelet[2982]: E0121 05:54:42.233423 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.234612 kubelet[2982]: W0121 05:54:42.233455 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.234612 kubelet[2982]: E0121 05:54:42.233484 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.237976 kubelet[2982]: E0121 05:54:42.237532 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.237976 kubelet[2982]: W0121 05:54:42.237563 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.237976 kubelet[2982]: E0121 05:54:42.237592 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.238522 kubelet[2982]: E0121 05:54:42.238228 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.238522 kubelet[2982]: W0121 05:54:42.238418 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.238522 kubelet[2982]: E0121 05:54:42.238441 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.243565 kubelet[2982]: E0121 05:54:42.243474 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.243565 kubelet[2982]: W0121 05:54:42.243494 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.243565 kubelet[2982]: E0121 05:54:42.243510 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.244916 kubelet[2982]: E0121 05:54:42.244244 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.244916 kubelet[2982]: W0121 05:54:42.244256 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.244916 kubelet[2982]: E0121 05:54:42.244269 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.247375 kubelet[2982]: E0121 05:54:42.244945 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.247375 kubelet[2982]: W0121 05:54:42.244957 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.247375 kubelet[2982]: E0121 05:54:42.244969 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.247375 kubelet[2982]: E0121 05:54:42.245352 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.247375 kubelet[2982]: W0121 05:54:42.245364 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.247375 kubelet[2982]: E0121 05:54:42.245376 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.247375 kubelet[2982]: E0121 05:54:42.246137 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.247375 kubelet[2982]: W0121 05:54:42.246154 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.247375 kubelet[2982]: E0121 05:54:42.246169 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.249450 kubelet[2982]: E0121 05:54:42.248192 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.249450 kubelet[2982]: W0121 05:54:42.248206 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.249450 kubelet[2982]: E0121 05:54:42.248778 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.254973 kubelet[2982]: E0121 05:54:42.254372 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.254973 kubelet[2982]: W0121 05:54:42.254395 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.254973 kubelet[2982]: E0121 05:54:42.254413 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.258153 kubelet[2982]: E0121 05:54:42.256866 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.258153 kubelet[2982]: W0121 05:54:42.256942 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.258153 kubelet[2982]: E0121 05:54:42.256961 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.258153 kubelet[2982]: E0121 05:54:42.257628 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.259900 kubelet[2982]: W0121 05:54:42.258908 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.259900 kubelet[2982]: E0121 05:54:42.259159 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.260768 kubelet[2982]: E0121 05:54:42.260489 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.260768 kubelet[2982]: W0121 05:54:42.260506 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.260768 kubelet[2982]: E0121 05:54:42.260522 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.263759 kubelet[2982]: E0121 05:54:42.263586 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.263759 kubelet[2982]: W0121 05:54:42.263603 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.263759 kubelet[2982]: E0121 05:54:42.263618 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.269250 kubelet[2982]: E0121 05:54:42.268559 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.269250 kubelet[2982]: W0121 05:54:42.268578 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.269250 kubelet[2982]: E0121 05:54:42.268595 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.276116 kubelet[2982]: E0121 05:54:42.276012 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.276257 kubelet[2982]: W0121 05:54:42.276238 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.276337 kubelet[2982]: E0121 05:54:42.276322 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.282911 kubelet[2982]: E0121 05:54:42.282886 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.283156 kubelet[2982]: W0121 05:54:42.283134 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.284402 kubelet[2982]: E0121 05:54:42.284381 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.288857 kubelet[2982]: E0121 05:54:42.288501 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.288857 kubelet[2982]: W0121 05:54:42.288522 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.288857 kubelet[2982]: E0121 05:54:42.288540 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.294519 kubelet[2982]: E0121 05:54:42.293892 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.294519 kubelet[2982]: W0121 05:54:42.293915 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.294519 kubelet[2982]: E0121 05:54:42.293932 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.301564 kubelet[2982]: E0121 05:54:42.301524 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.302202 kubelet[2982]: W0121 05:54:42.302174 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.302314 kubelet[2982]: E0121 05:54:42.302295 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.309191 kubelet[2982]: E0121 05:54:42.302980 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.309191 kubelet[2982]: W0121 05:54:42.302996 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.309191 kubelet[2982]: E0121 05:54:42.303011 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.320562 kubelet[2982]: E0121 05:54:42.320521 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.320971 kubelet[2982]: W0121 05:54:42.320945 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.323456 kubelet[2982]: E0121 05:54:42.323427 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.327459 kubelet[2982]: E0121 05:54:42.327438 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.327595 kubelet[2982]: W0121 05:54:42.327572 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.327822 kubelet[2982]: E0121 05:54:42.327802 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.329981 kubelet[2982]: E0121 05:54:42.329957 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.330186 kubelet[2982]: W0121 05:54:42.330167 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.330274 kubelet[2982]: E0121 05:54:42.330255 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.338346 kubelet[2982]: E0121 05:54:42.338320 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.338495 kubelet[2982]: W0121 05:54:42.338475 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.338585 kubelet[2982]: E0121 05:54:42.338568 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.342554 kubelet[2982]: E0121 05:54:42.342535 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.342841 kubelet[2982]: W0121 05:54:42.342822 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.342937 kubelet[2982]: E0121 05:54:42.342920 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.346933 kubelet[2982]: E0121 05:54:42.346913 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.347210 kubelet[2982]: W0121 05:54:42.347187 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.348585 kubelet[2982]: E0121 05:54:42.348564 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.349594 kubelet[2982]: E0121 05:54:42.349575 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.349817 kubelet[2982]: W0121 05:54:42.349799 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.349906 kubelet[2982]: E0121 05:54:42.349890 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.352599 kubelet[2982]: E0121 05:54:42.352577 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.352968 kubelet[2982]: W0121 05:54:42.352836 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.353546 kubelet[2982]: E0121 05:54:42.353525 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.354280 kubelet[2982]: E0121 05:54:42.354262 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.354377 kubelet[2982]: W0121 05:54:42.354359 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.354461 kubelet[2982]: E0121 05:54:42.354445 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.355372 kubelet[2982]: E0121 05:54:42.355352 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.355489 kubelet[2982]: W0121 05:54:42.355467 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.355589 kubelet[2982]: E0121 05:54:42.355569 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.361861 kubelet[2982]: E0121 05:54:42.361836 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.362019 kubelet[2982]: W0121 05:54:42.361999 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.362183 kubelet[2982]: E0121 05:54:42.362166 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.365020 kubelet[2982]: E0121 05:54:42.364934 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.365297 kubelet[2982]: W0121 05:54:42.365207 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.365404 kubelet[2982]: E0121 05:54:42.365384 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.367136 kubelet[2982]: E0121 05:54:42.367111 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.372890 kubelet[2982]: W0121 05:54:42.372757 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.372890 kubelet[2982]: E0121 05:54:42.372792 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.397013 kubelet[2982]: E0121 05:54:42.396842 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.397013 kubelet[2982]: W0121 05:54:42.397021 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.398472 kubelet[2982]: E0121 05:54:42.397258 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.401457 kubelet[2982]: E0121 05:54:42.400801 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.401457 kubelet[2982]: W0121 05:54:42.400848 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.401457 kubelet[2982]: E0121 05:54:42.400886 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.403529 kubelet[2982]: E0121 05:54:42.402001 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.403529 kubelet[2982]: W0121 05:54:42.402235 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.403529 kubelet[2982]: E0121 05:54:42.402256 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.408511 kubelet[2982]: E0121 05:54:42.408299 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.408511 kubelet[2982]: W0121 05:54:42.408329 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.408511 kubelet[2982]: E0121 05:54:42.408353 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.410621 kubelet[2982]: E0121 05:54:42.410137 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.410621 kubelet[2982]: W0121 05:54:42.410162 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.410621 kubelet[2982]: E0121 05:54:42.410177 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:42.411723 kubelet[2982]: E0121 05:54:42.411322 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:42.411723 kubelet[2982]: W0121 05:54:42.411337 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:42.411723 kubelet[2982]: E0121 05:54:42.411351 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:43.236390 kubelet[2982]: E0121 05:54:43.235879 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:44.212502 kubelet[2982]: E0121 05:54:44.210358 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:44.318874 kubelet[2982]: E0121 05:54:44.313911 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:44.318874 kubelet[2982]: W0121 05:54:44.314013 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:44.318874 kubelet[2982]: E0121 05:54:44.314048 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:44.318874 kubelet[2982]: E0121 05:54:44.318462 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:44.318874 kubelet[2982]: W0121 05:54:44.318475 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:44.318874 kubelet[2982]: E0121 05:54:44.318493 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:44.326314 kubelet[2982]: E0121 05:54:44.326100 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:44.326314 kubelet[2982]: W0121 05:54:44.326214 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:44.326594 kubelet[2982]: E0121 05:54:44.326487 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:44.330344 kubelet[2982]: E0121 05:54:44.330220 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:44.330344 kubelet[2982]: W0121 05:54:44.330339 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:44.330469 kubelet[2982]: E0121 05:54:44.330362 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:44.332114 kubelet[2982]: E0121 05:54:44.332071 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:44.332114 kubelet[2982]: W0121 05:54:44.332094 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:44.332114 kubelet[2982]: E0121 05:54:44.332113 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:45.213724 kubelet[2982]: E0121 05:54:45.210169 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:47.213896 kubelet[2982]: E0121 05:54:47.211457 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:49.215625 kubelet[2982]: E0121 05:54:49.209472 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:50.846273 containerd[1639]: time="2026-01-21T05:54:50.842841598Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:54:50.850996 containerd[1639]: time="2026-01-21T05:54:50.850606701Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=33735893" Jan 21 05:54:50.860625 containerd[1639]: time="2026-01-21T05:54:50.859600824Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:54:50.895047 containerd[1639]: time="2026-01-21T05:54:50.894308429Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:54:50.900182 containerd[1639]: time="2026-01-21T05:54:50.896383047Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 15.168175992s" Jan 21 05:54:50.900182 containerd[1639]: time="2026-01-21T05:54:50.896430309Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Jan 21 05:54:50.930439 containerd[1639]: time="2026-01-21T05:54:50.929143677Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Jan 21 05:54:51.011599 containerd[1639]: time="2026-01-21T05:54:51.011348292Z" level=info msg="CreateContainer within sandbox \"a812ac7a46cdd6272a729c7bf8355e2864a58adb68f334d1825155c4e53dd22e\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jan 21 05:54:51.073455 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3564702605.mount: Deactivated successfully. Jan 21 05:54:51.089015 containerd[1639]: time="2026-01-21T05:54:51.087407522Z" level=info msg="Container fcdfc4c2ca702a59181a3e1ce931c8b117eded1577d675bff313c3628d2c3335: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:54:51.175879 containerd[1639]: time="2026-01-21T05:54:51.175282725Z" level=info msg="CreateContainer within sandbox \"a812ac7a46cdd6272a729c7bf8355e2864a58adb68f334d1825155c4e53dd22e\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"fcdfc4c2ca702a59181a3e1ce931c8b117eded1577d675bff313c3628d2c3335\"" Jan 21 05:54:51.187428 containerd[1639]: time="2026-01-21T05:54:51.187299201Z" level=info msg="StartContainer for \"fcdfc4c2ca702a59181a3e1ce931c8b117eded1577d675bff313c3628d2c3335\"" Jan 21 05:54:51.199358 containerd[1639]: time="2026-01-21T05:54:51.196448274Z" level=info msg="connecting to shim fcdfc4c2ca702a59181a3e1ce931c8b117eded1577d675bff313c3628d2c3335" address="unix:///run/containerd/s/d790881195b101fa9e356752fea53ebfb66421b00a10d1ccc5a1707fcc70c91f" protocol=ttrpc version=3 Jan 21 05:54:51.209292 kubelet[2982]: E0121 05:54:51.209108 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:51.388020 systemd[1]: Started cri-containerd-fcdfc4c2ca702a59181a3e1ce931c8b117eded1577d675bff313c3628d2c3335.scope - libcontainer container fcdfc4c2ca702a59181a3e1ce931c8b117eded1577d675bff313c3628d2c3335. Jan 21 05:54:51.727382 kernel: kauditd_printk_skb: 52 callbacks suppressed Jan 21 05:54:51.727584 kernel: audit: type=1334 audit(1768974891.714:553): prog-id=159 op=LOAD Jan 21 05:54:51.714000 audit: BPF prog-id=159 op=LOAD Jan 21 05:54:51.728000 audit: BPF prog-id=160 op=LOAD Jan 21 05:54:51.774335 kernel: audit: type=1334 audit(1768974891.728:554): prog-id=160 op=LOAD Jan 21 05:54:51.774497 kernel: audit: type=1300 audit(1768974891.728:554): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c238 a2=98 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:51.728000 audit[3679]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c238 a2=98 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:51.728000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:51.806256 kernel: audit: type=1327 audit(1768974891.728:554): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:51.728000 audit: BPF prog-id=160 op=UNLOAD Jan 21 05:54:51.728000 audit[3679]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:51.878388 kernel: audit: type=1334 audit(1768974891.728:555): prog-id=160 op=UNLOAD Jan 21 05:54:51.878540 kernel: audit: type=1300 audit(1768974891.728:555): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:51.728000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:51.743000 audit: BPF prog-id=161 op=LOAD Jan 21 05:54:51.955344 kernel: audit: type=1327 audit(1768974891.728:555): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:51.955465 kernel: audit: type=1334 audit(1768974891.743:556): prog-id=161 op=LOAD Jan 21 05:54:51.743000 audit[3679]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c488 a2=98 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:52.016753 kernel: audit: type=1300 audit(1768974891.743:556): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c488 a2=98 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:51.743000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:52.090479 kernel: audit: type=1327 audit(1768974891.743:556): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:51.743000 audit: BPF prog-id=162 op=LOAD Jan 21 05:54:51.743000 audit[3679]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00010c218 a2=98 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:51.743000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:51.743000 audit: BPF prog-id=162 op=UNLOAD Jan 21 05:54:51.743000 audit[3679]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:51.743000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:51.743000 audit: BPF prog-id=161 op=UNLOAD Jan 21 05:54:51.743000 audit[3679]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:51.743000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:51.743000 audit: BPF prog-id=163 op=LOAD Jan 21 05:54:51.743000 audit[3679]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c6e8 a2=98 a3=0 items=0 ppid=3436 pid=3679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:51.743000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6663646663346332636137303261353931383161336531636539333163 Jan 21 05:54:52.429359 containerd[1639]: time="2026-01-21T05:54:52.422136220Z" level=info msg="StartContainer for \"fcdfc4c2ca702a59181a3e1ce931c8b117eded1577d675bff313c3628d2c3335\" returns successfully" Jan 21 05:54:52.704422 kubelet[2982]: E0121 05:54:52.699141 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:52.780438 kubelet[2982]: E0121 05:54:52.780357 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.780438 kubelet[2982]: W0121 05:54:52.780394 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.780438 kubelet[2982]: E0121 05:54:52.780424 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.791114 kubelet[2982]: E0121 05:54:52.790436 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.791114 kubelet[2982]: W0121 05:54:52.790486 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.791114 kubelet[2982]: E0121 05:54:52.790520 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.798262 kubelet[2982]: E0121 05:54:52.797868 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.798262 kubelet[2982]: W0121 05:54:52.797962 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.798262 kubelet[2982]: E0121 05:54:52.797995 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.801519 kubelet[2982]: E0121 05:54:52.801316 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.801519 kubelet[2982]: W0121 05:54:52.801338 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.801519 kubelet[2982]: E0121 05:54:52.801360 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.819953 kubelet[2982]: E0121 05:54:52.817883 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.819953 kubelet[2982]: W0121 05:54:52.817917 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.819953 kubelet[2982]: E0121 05:54:52.817950 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.826307 kubelet[2982]: E0121 05:54:52.826011 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.826307 kubelet[2982]: W0121 05:54:52.826035 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.826307 kubelet[2982]: E0121 05:54:52.826259 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.835776 kubelet[2982]: E0121 05:54:52.835305 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.835776 kubelet[2982]: W0121 05:54:52.835335 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.835776 kubelet[2982]: E0121 05:54:52.835365 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.837786 kubelet[2982]: E0121 05:54:52.837603 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.838373 kubelet[2982]: W0121 05:54:52.838232 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.838373 kubelet[2982]: E0121 05:54:52.838336 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.855374 kubelet[2982]: E0121 05:54:52.854945 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.855374 kubelet[2982]: W0121 05:54:52.855126 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.855374 kubelet[2982]: E0121 05:54:52.855167 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.856900 kubelet[2982]: E0121 05:54:52.856858 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.856900 kubelet[2982]: W0121 05:54:52.856881 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.857015 kubelet[2982]: E0121 05:54:52.856905 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.866451 kubelet[2982]: E0121 05:54:52.865552 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.866451 kubelet[2982]: W0121 05:54:52.865841 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.866451 kubelet[2982]: E0121 05:54:52.865878 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.878305 kubelet[2982]: E0121 05:54:52.871261 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.878305 kubelet[2982]: W0121 05:54:52.871286 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.878305 kubelet[2982]: E0121 05:54:52.871314 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.878305 kubelet[2982]: E0121 05:54:52.872919 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.878305 kubelet[2982]: W0121 05:54:52.872938 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.878305 kubelet[2982]: E0121 05:54:52.872962 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.892538 kubelet[2982]: E0121 05:54:52.884042 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.892538 kubelet[2982]: W0121 05:54:52.892198 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.892538 kubelet[2982]: E0121 05:54:52.892231 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.900293 kubelet[2982]: E0121 05:54:52.900012 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.904251 kubelet[2982]: W0121 05:54:52.900040 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.904251 kubelet[2982]: E0121 05:54:52.902247 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.908001 kubelet[2982]: E0121 05:54:52.907867 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.908001 kubelet[2982]: W0121 05:54:52.907892 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.908001 kubelet[2982]: E0121 05:54:52.907920 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.912846 kubelet[2982]: E0121 05:54:52.911897 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.912846 kubelet[2982]: W0121 05:54:52.911922 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.912846 kubelet[2982]: E0121 05:54:52.911944 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.921785 kubelet[2982]: E0121 05:54:52.921177 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.921785 kubelet[2982]: W0121 05:54:52.921203 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.921785 kubelet[2982]: E0121 05:54:52.921229 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.924221 kubelet[2982]: E0121 05:54:52.923970 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.924221 kubelet[2982]: W0121 05:54:52.924133 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.924221 kubelet[2982]: E0121 05:54:52.924159 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.943797 kubelet[2982]: E0121 05:54:52.943423 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.943797 kubelet[2982]: W0121 05:54:52.943463 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.943797 kubelet[2982]: E0121 05:54:52.943577 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.946809 kubelet[2982]: E0121 05:54:52.946370 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.948883 kubelet[2982]: W0121 05:54:52.948605 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.951917 kubelet[2982]: E0121 05:54:52.951512 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:52.967451 kubelet[2982]: E0121 05:54:52.965540 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:52.969749 kubelet[2982]: W0121 05:54:52.967795 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:52.969749 kubelet[2982]: E0121 05:54:52.967847 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.003144 kubelet[2982]: E0121 05:54:53.001574 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.003144 kubelet[2982]: W0121 05:54:53.002578 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.003144 kubelet[2982]: E0121 05:54:53.002619 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.006850 kubelet[2982]: E0121 05:54:53.006353 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.006850 kubelet[2982]: W0121 05:54:53.006376 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.006850 kubelet[2982]: E0121 05:54:53.006400 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.011025 kubelet[2982]: E0121 05:54:53.010616 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.012536 kubelet[2982]: W0121 05:54:53.012021 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.013847 kubelet[2982]: E0121 05:54:53.013536 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.021168 kubelet[2982]: E0121 05:54:53.020622 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.022038 kubelet[2982]: W0121 05:54:53.022010 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.022780 kubelet[2982]: E0121 05:54:53.022205 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.024204 kubelet[2982]: E0121 05:54:53.024183 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.024500 kubelet[2982]: W0121 05:54:53.024480 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.024601 kubelet[2982]: E0121 05:54:53.024581 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.030022 kubelet[2982]: E0121 05:54:53.029626 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.030744 kubelet[2982]: W0121 05:54:53.030718 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.031262 kubelet[2982]: E0121 05:54:53.030919 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.037491 kubelet[2982]: E0121 05:54:53.037163 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.037491 kubelet[2982]: W0121 05:54:53.037187 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.037491 kubelet[2982]: E0121 05:54:53.037209 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.041986 kubelet[2982]: E0121 05:54:53.039792 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.041986 kubelet[2982]: W0121 05:54:53.039812 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.041986 kubelet[2982]: E0121 05:54:53.039832 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.046145 kubelet[2982]: E0121 05:54:53.045928 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.046145 kubelet[2982]: W0121 05:54:53.045950 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.046145 kubelet[2982]: E0121 05:54:53.045970 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.054452 kubelet[2982]: E0121 05:54:53.054208 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.054452 kubelet[2982]: W0121 05:54:53.054242 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.054452 kubelet[2982]: E0121 05:54:53.054277 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.063849 kubelet[2982]: E0121 05:54:53.058558 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.063849 kubelet[2982]: W0121 05:54:53.060856 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.063849 kubelet[2982]: E0121 05:54:53.060901 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.135365 containerd[1639]: time="2026-01-21T05:54:53.133248620Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:54:53.143588 containerd[1639]: time="2026-01-21T05:54:53.140565905Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=0" Jan 21 05:54:53.143588 containerd[1639]: time="2026-01-21T05:54:53.142976989Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:54:53.150591 containerd[1639]: time="2026-01-21T05:54:53.150206606Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:54:53.176472 containerd[1639]: time="2026-01-21T05:54:53.175387409Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 2.246089691s" Jan 21 05:54:53.176472 containerd[1639]: time="2026-01-21T05:54:53.175434711Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Jan 21 05:54:53.288033 containerd[1639]: time="2026-01-21T05:54:53.273373495Z" level=info msg="CreateContainer within sandbox \"e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jan 21 05:54:53.289052 kubelet[2982]: E0121 05:54:53.288929 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:53.382504 containerd[1639]: time="2026-01-21T05:54:53.382454698Z" level=info msg="Container 354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:54:53.394588 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2362445044.mount: Deactivated successfully. Jan 21 05:54:53.429286 containerd[1639]: time="2026-01-21T05:54:53.428829396Z" level=info msg="CreateContainer within sandbox \"e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc\"" Jan 21 05:54:53.430230 containerd[1639]: time="2026-01-21T05:54:53.430202237Z" level=info msg="StartContainer for \"354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc\"" Jan 21 05:54:53.435848 containerd[1639]: time="2026-01-21T05:54:53.435605705Z" level=info msg="connecting to shim 354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc" address="unix:///run/containerd/s/a097f43dd299e2930c4a70b814880e743e86b92e44c5686bf5acf59179f1242e" protocol=ttrpc version=3 Jan 21 05:54:53.655064 systemd[1]: Started cri-containerd-354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc.scope - libcontainer container 354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc. Jan 21 05:54:53.739588 kubelet[2982]: E0121 05:54:53.739212 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:53.791489 kubelet[2982]: E0121 05:54:53.791066 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.792295 kubelet[2982]: W0121 05:54:53.791816 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.793218 kubelet[2982]: E0121 05:54:53.793021 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.799465 kubelet[2982]: E0121 05:54:53.798947 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.799465 kubelet[2982]: W0121 05:54:53.798974 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.799465 kubelet[2982]: E0121 05:54:53.799002 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.802786 kubelet[2982]: E0121 05:54:53.801916 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.802910 kubelet[2982]: W0121 05:54:53.802888 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.803014 kubelet[2982]: E0121 05:54:53.802993 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.805246 kubelet[2982]: E0121 05:54:53.805220 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.805337 kubelet[2982]: W0121 05:54:53.805320 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.806482 kubelet[2982]: E0121 05:54:53.806461 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.810498 kubelet[2982]: E0121 05:54:53.808609 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.810611 kubelet[2982]: W0121 05:54:53.810590 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.810885 kubelet[2982]: E0121 05:54:53.810864 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.816797 kubelet[2982]: E0121 05:54:53.815847 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.816797 kubelet[2982]: W0121 05:54:53.815875 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.816797 kubelet[2982]: E0121 05:54:53.815896 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.818899 kubelet[2982]: E0121 05:54:53.818878 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.818997 kubelet[2982]: W0121 05:54:53.818980 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.819109 kubelet[2982]: E0121 05:54:53.819087 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.822250 kubelet[2982]: E0121 05:54:53.822220 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.822364 kubelet[2982]: W0121 05:54:53.822345 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.823077 kubelet[2982]: E0121 05:54:53.822977 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.825814 kubelet[2982]: E0121 05:54:53.825089 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.825814 kubelet[2982]: W0121 05:54:53.825113 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.825814 kubelet[2982]: E0121 05:54:53.825220 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.828082 kubelet[2982]: E0121 05:54:53.828060 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.828342 kubelet[2982]: W0121 05:54:53.828321 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.829530 kubelet[2982]: E0121 05:54:53.829507 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.834095 kubelet[2982]: E0121 05:54:53.834067 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.834343 kubelet[2982]: W0121 05:54:53.834321 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.834457 kubelet[2982]: E0121 05:54:53.834436 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.836537 kubelet[2982]: E0121 05:54:53.836475 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.836537 kubelet[2982]: W0121 05:54:53.836494 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.836537 kubelet[2982]: E0121 05:54:53.836510 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.839361 kubelet[2982]: E0121 05:54:53.839038 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.839361 kubelet[2982]: W0121 05:54:53.839057 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.839361 kubelet[2982]: E0121 05:54:53.839073 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.841061 kubelet[2982]: E0121 05:54:53.840806 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.841061 kubelet[2982]: W0121 05:54:53.840830 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.841061 kubelet[2982]: E0121 05:54:53.840848 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.841854 kubelet[2982]: I0121 05:54:53.841791 2982 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-5596648d99-m7djd" podStartSLOduration=7.656281273 podStartE2EDuration="22.841620664s" podCreationTimestamp="2026-01-21 05:54:31 +0000 UTC" firstStartedPulling="2026-01-21 05:54:35.727516614 +0000 UTC m=+69.937166317" lastFinishedPulling="2026-01-21 05:54:50.912856004 +0000 UTC m=+85.122505708" observedRunningTime="2026-01-21 05:54:52.818180049 +0000 UTC m=+87.027829752" watchObservedRunningTime="2026-01-21 05:54:53.841620664 +0000 UTC m=+88.051270386" Jan 21 05:54:53.845038 kubelet[2982]: E0121 05:54:53.844574 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.845038 kubelet[2982]: W0121 05:54:53.844594 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.845038 kubelet[2982]: E0121 05:54:53.844611 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.929000 audit[3789]: NETFILTER_CFG table=filter:119 family=2 entries=21 op=nft_register_rule pid=3789 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:53.929000 audit[3789]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc90e5f3f0 a2=0 a3=7ffc90e5f3dc items=0 ppid=3144 pid=3789 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:53.929000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:53.939756 kubelet[2982]: E0121 05:54:53.939311 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.939756 kubelet[2982]: W0121 05:54:53.939421 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.939756 kubelet[2982]: E0121 05:54:53.939450 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.941355 kubelet[2982]: E0121 05:54:53.940849 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.943387 kubelet[2982]: W0121 05:54:53.941518 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.943387 kubelet[2982]: E0121 05:54:53.941616 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.949933 kubelet[2982]: E0121 05:54:53.944783 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.949933 kubelet[2982]: W0121 05:54:53.944805 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.949933 kubelet[2982]: E0121 05:54:53.944822 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.949933 kubelet[2982]: E0121 05:54:53.945820 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.949933 kubelet[2982]: W0121 05:54:53.945833 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.949933 kubelet[2982]: E0121 05:54:53.945846 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.949933 kubelet[2982]: E0121 05:54:53.946578 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.949933 kubelet[2982]: W0121 05:54:53.946591 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.949933 kubelet[2982]: E0121 05:54:53.946603 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.953129 kubelet[2982]: E0121 05:54:53.951800 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.953129 kubelet[2982]: W0121 05:54:53.951938 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.953129 kubelet[2982]: E0121 05:54:53.951961 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.959126 kubelet[2982]: E0121 05:54:53.954824 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.959126 kubelet[2982]: W0121 05:54:53.954848 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.959126 kubelet[2982]: E0121 05:54:53.954868 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.959126 kubelet[2982]: E0121 05:54:53.957261 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.959126 kubelet[2982]: W0121 05:54:53.957275 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.959126 kubelet[2982]: E0121 05:54:53.957294 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.967406 kubelet[2982]: E0121 05:54:53.967003 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.967406 kubelet[2982]: W0121 05:54:53.967038 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.967406 kubelet[2982]: E0121 05:54:53.967073 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.966000 audit[3789]: NETFILTER_CFG table=nat:120 family=2 entries=19 op=nft_register_chain pid=3789 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:54:53.969557 kubelet[2982]: E0121 05:54:53.969540 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.969815 kubelet[2982]: W0121 05:54:53.969627 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.971567 kubelet[2982]: E0121 05:54:53.971543 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.966000 audit[3789]: SYSCALL arch=c000003e syscall=46 success=yes exit=6276 a0=3 a1=7ffc90e5f3f0 a2=0 a3=7ffc90e5f3dc items=0 ppid=3144 pid=3789 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:53.966000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:54:53.974127 kubelet[2982]: E0121 05:54:53.973869 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.974127 kubelet[2982]: W0121 05:54:53.973967 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.974127 kubelet[2982]: E0121 05:54:53.973997 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.982589 kubelet[2982]: E0121 05:54:53.982293 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.982589 kubelet[2982]: W0121 05:54:53.982405 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.982589 kubelet[2982]: E0121 05:54:53.982440 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.985265 kubelet[2982]: E0121 05:54:53.985239 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.985354 kubelet[2982]: W0121 05:54:53.985336 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.985437 kubelet[2982]: E0121 05:54:53.985420 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.988043 kubelet[2982]: E0121 05:54:53.987801 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.988043 kubelet[2982]: W0121 05:54:53.987820 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.988043 kubelet[2982]: E0121 05:54:53.987840 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.993944 kubelet[2982]: E0121 05:54:53.993849 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.994055 kubelet[2982]: W0121 05:54:53.994034 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.994143 kubelet[2982]: E0121 05:54:53.994125 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.996787 kubelet[2982]: E0121 05:54:53.996393 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.996787 kubelet[2982]: W0121 05:54:53.996411 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.996787 kubelet[2982]: E0121 05:54:53.996426 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.998117 kubelet[2982]: E0121 05:54:53.998096 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.998295 kubelet[2982]: W0121 05:54:53.998277 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.998375 kubelet[2982]: E0121 05:54:53.998359 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:53.999521 kubelet[2982]: E0121 05:54:53.999501 2982 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 21 05:54:53.999609 kubelet[2982]: W0121 05:54:53.999589 2982 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 21 05:54:53.999853 kubelet[2982]: E0121 05:54:53.999836 2982 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 21 05:54:54.146000 audit: BPF prog-id=164 op=LOAD Jan 21 05:54:54.146000 audit[3755]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=3542 pid=3755 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:54.146000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3335346436366661373533363236353165636163653763303864333837 Jan 21 05:54:54.146000 audit: BPF prog-id=165 op=LOAD Jan 21 05:54:54.146000 audit[3755]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=3542 pid=3755 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:54.146000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3335346436366661373533363236353165636163653763303864333837 Jan 21 05:54:54.146000 audit: BPF prog-id=165 op=UNLOAD Jan 21 05:54:54.146000 audit[3755]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=3755 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:54.146000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3335346436366661373533363236353165636163653763303864333837 Jan 21 05:54:54.146000 audit: BPF prog-id=164 op=UNLOAD Jan 21 05:54:54.146000 audit[3755]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=3755 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:54.146000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3335346436366661373533363236353165636163653763303864333837 Jan 21 05:54:54.147000 audit: BPF prog-id=166 op=LOAD Jan 21 05:54:54.147000 audit[3755]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001866e8 a2=98 a3=0 items=0 ppid=3542 pid=3755 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:54:54.147000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3335346436366661373533363236353165636163653763303864333837 Jan 21 05:54:54.309378 containerd[1639]: time="2026-01-21T05:54:54.308905551Z" level=info msg="StartContainer for \"354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc\" returns successfully" Jan 21 05:54:54.384130 systemd[1]: cri-containerd-354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc.scope: Deactivated successfully. Jan 21 05:54:54.396000 audit: BPF prog-id=166 op=UNLOAD Jan 21 05:54:54.428861 containerd[1639]: time="2026-01-21T05:54:54.425083397Z" level=info msg="received container exit event container_id:\"354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc\" id:\"354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc\" pid:3769 exited_at:{seconds:1768974894 nanos:409418807}" Jan 21 05:54:54.850579 kubelet[2982]: E0121 05:54:54.840843 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:54.861959 kubelet[2982]: E0121 05:54:54.855906 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:55.039971 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-354d66fa75362651ecace7c08d3877c57f696ede6a2d792ccd11c90dbedc60fc-rootfs.mount: Deactivated successfully. Jan 21 05:54:55.214982 kubelet[2982]: E0121 05:54:55.213052 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:55.214982 kubelet[2982]: E0121 05:54:55.214076 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:55.907543 kubelet[2982]: E0121 05:54:55.903586 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:54:55.914778 containerd[1639]: time="2026-01-21T05:54:55.914593738Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Jan 21 05:54:57.257348 kubelet[2982]: E0121 05:54:57.256279 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:54:59.210993 kubelet[2982]: E0121 05:54:59.210101 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:01.216328 kubelet[2982]: E0121 05:55:01.216191 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:03.210967 kubelet[2982]: E0121 05:55:03.208055 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:05.211523 kubelet[2982]: E0121 05:55:05.211279 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:07.212813 kubelet[2982]: E0121 05:55:07.211376 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:09.209562 kubelet[2982]: E0121 05:55:09.209129 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:10.304061 containerd[1639]: time="2026-01-21T05:55:10.303894077Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:55:10.311578 containerd[1639]: time="2026-01-21T05:55:10.311241793Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70442291" Jan 21 05:55:10.321291 containerd[1639]: time="2026-01-21T05:55:10.321217742Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:55:10.329992 containerd[1639]: time="2026-01-21T05:55:10.329323184Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:55:10.331148 containerd[1639]: time="2026-01-21T05:55:10.330956533Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 14.416318919s" Jan 21 05:55:10.331148 containerd[1639]: time="2026-01-21T05:55:10.331072857Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Jan 21 05:55:10.346878 containerd[1639]: time="2026-01-21T05:55:10.346276821Z" level=info msg="CreateContainer within sandbox \"e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 21 05:55:10.383029 containerd[1639]: time="2026-01-21T05:55:10.382600040Z" level=info msg="Container e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:55:10.414493 containerd[1639]: time="2026-01-21T05:55:10.411188822Z" level=info msg="CreateContainer within sandbox \"e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409\"" Jan 21 05:55:10.417082 containerd[1639]: time="2026-01-21T05:55:10.417047855Z" level=info msg="StartContainer for \"e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409\"" Jan 21 05:55:10.425509 containerd[1639]: time="2026-01-21T05:55:10.425320615Z" level=info msg="connecting to shim e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409" address="unix:///run/containerd/s/a097f43dd299e2930c4a70b814880e743e86b92e44c5686bf5acf59179f1242e" protocol=ttrpc version=3 Jan 21 05:55:10.555207 systemd[1]: Started cri-containerd-e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409.scope - libcontainer container e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409. Jan 21 05:55:10.741000 audit: BPF prog-id=167 op=LOAD Jan 21 05:55:10.756790 kernel: kauditd_printk_skb: 34 callbacks suppressed Jan 21 05:55:10.756947 kernel: audit: type=1334 audit(1768974910.741:569): prog-id=167 op=LOAD Jan 21 05:55:10.741000 audit[3856]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3542 pid=3856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:10.792076 kernel: audit: type=1300 audit(1768974910.741:569): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3542 pid=3856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:10.792914 kernel: audit: type=1327 audit(1768974910.741:569): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538303162623764313530306461653866663135346339393437383266 Jan 21 05:55:10.741000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538303162623764313530306461653866663135346339393437383266 Jan 21 05:55:10.746000 audit: BPF prog-id=168 op=LOAD Jan 21 05:55:10.746000 audit[3856]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=3542 pid=3856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:10.832806 kernel: audit: type=1334 audit(1768974910.746:570): prog-id=168 op=LOAD Jan 21 05:55:10.832876 kernel: audit: type=1300 audit(1768974910.746:570): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=3542 pid=3856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:10.746000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538303162623764313530306461653866663135346339393437383266 Jan 21 05:55:10.748000 audit: BPF prog-id=168 op=UNLOAD Jan 21 05:55:10.972583 kernel: audit: type=1327 audit(1768974910.746:570): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538303162623764313530306461653866663135346339393437383266 Jan 21 05:55:10.973793 kernel: audit: type=1334 audit(1768974910.748:571): prog-id=168 op=UNLOAD Jan 21 05:55:10.975252 kernel: audit: type=1300 audit(1768974910.748:571): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=3856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:10.748000 audit[3856]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=3856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:11.027169 kernel: audit: type=1327 audit(1768974910.748:571): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538303162623764313530306461653866663135346339393437383266 Jan 21 05:55:10.748000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538303162623764313530306461653866663135346339393437383266 Jan 21 05:55:11.070520 containerd[1639]: time="2026-01-21T05:55:11.066847846Z" level=info msg="StartContainer for \"e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409\" returns successfully" Jan 21 05:55:10.748000 audit: BPF prog-id=167 op=UNLOAD Jan 21 05:55:11.112076 kernel: audit: type=1334 audit(1768974910.748:572): prog-id=167 op=UNLOAD Jan 21 05:55:10.748000 audit[3856]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=3856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:10.748000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538303162623764313530306461653866663135346339393437383266 Jan 21 05:55:10.748000 audit: BPF prog-id=169 op=LOAD Jan 21 05:55:10.748000 audit[3856]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=3542 pid=3856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:10.748000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538303162623764313530306461653866663135346339393437383266 Jan 21 05:55:11.198377 kubelet[2982]: E0121 05:55:11.198221 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:11.215217 kubelet[2982]: E0121 05:55:11.212945 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:12.204966 kubelet[2982]: E0121 05:55:12.203808 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:13.209532 kubelet[2982]: E0121 05:55:13.209414 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:15.232026 kubelet[2982]: E0121 05:55:15.228385 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:16.528543 kubelet[2982]: E0121 05:55:16.526087 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:16.581165 systemd[1]: cri-containerd-e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409.scope: Deactivated successfully. Jan 21 05:55:16.582021 systemd[1]: cri-containerd-e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409.scope: Consumed 3.200s CPU time, 176.7M memory peak, 4.3M read from disk, 171.3M written to disk. Jan 21 05:55:16.603759 kernel: kauditd_printk_skb: 5 callbacks suppressed Jan 21 05:55:16.604050 kernel: audit: type=1334 audit(1768974916.587:574): prog-id=169 op=UNLOAD Jan 21 05:55:16.587000 audit: BPF prog-id=169 op=UNLOAD Jan 21 05:55:16.604208 containerd[1639]: time="2026-01-21T05:55:16.601317531Z" level=info msg="received container exit event container_id:\"e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409\" id:\"e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409\" pid:3869 exited_at:{seconds:1768974916 nanos:600079689}" Jan 21 05:55:18.984056 kubelet[2982]: E0121 05:55:18.983500 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:23.790783 kubelet[2982]: I0121 05:55:23.774389 2982 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Jan 21 05:55:23.811847 kubelet[2982]: E0121 05:55:23.794308 2982 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.912s" Jan 21 05:55:23.847803 systemd[1]: Created slice kubepods-besteffort-pod487c6aa2_04b5_4684_9420_0e4b73799fd3.slice - libcontainer container kubepods-besteffort-pod487c6aa2_04b5_4684_9420_0e4b73799fd3.slice. Jan 21 05:55:23.886557 containerd[1639]: time="2026-01-21T05:55:23.885605378Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:24.321824 kubelet[2982]: I0121 05:55:24.286488 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/dbddb9a5-130c-465d-9677-8a7b919a18f8-calico-apiserver-certs\") pod \"calico-apiserver-fcddb958d-sjlxh\" (UID: \"dbddb9a5-130c-465d-9677-8a7b919a18f8\") " pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:55:24.503800 kubelet[2982]: I0121 05:55:24.502368 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wclgb\" (UniqueName: \"kubernetes.io/projected/dbddb9a5-130c-465d-9677-8a7b919a18f8-kube-api-access-wclgb\") pod \"calico-apiserver-fcddb958d-sjlxh\" (UID: \"dbddb9a5-130c-465d-9677-8a7b919a18f8\") " pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:55:24.503800 kubelet[2982]: I0121 05:55:24.502864 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5wjj\" (UniqueName: \"kubernetes.io/projected/dac6efad-1039-4242-b1c8-782ac61e5470-kube-api-access-n5wjj\") pod \"calico-apiserver-796779bb-f2fds\" (UID: \"dac6efad-1039-4242-b1c8-782ac61e5470\") " pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:55:24.503800 kubelet[2982]: I0121 05:55:24.503048 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw2vd\" (UniqueName: \"kubernetes.io/projected/36de057e-0cdd-42d8-a113-435fdf1d98e9-kube-api-access-rw2vd\") pod \"coredns-674b8bbfcf-sgzh7\" (UID: \"36de057e-0cdd-42d8-a113-435fdf1d98e9\") " pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:55:24.503800 kubelet[2982]: I0121 05:55:24.503086 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/59842dba-2f25-4ec6-86e9-919aa2afb77a-goldmane-key-pair\") pod \"goldmane-666569f655-m7pjj\" (UID: \"59842dba-2f25-4ec6-86e9-919aa2afb77a\") " pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:55:24.503800 kubelet[2982]: I0121 05:55:24.503111 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6x2t\" (UniqueName: \"kubernetes.io/projected/59842dba-2f25-4ec6-86e9-919aa2afb77a-kube-api-access-s6x2t\") pod \"goldmane-666569f655-m7pjj\" (UID: \"59842dba-2f25-4ec6-86e9-919aa2afb77a\") " pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:55:24.509345 kubelet[2982]: I0121 05:55:24.503159 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36de057e-0cdd-42d8-a113-435fdf1d98e9-config-volume\") pod \"coredns-674b8bbfcf-sgzh7\" (UID: \"36de057e-0cdd-42d8-a113-435fdf1d98e9\") " pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:55:24.509345 kubelet[2982]: I0121 05:55:24.503474 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59842dba-2f25-4ec6-86e9-919aa2afb77a-config\") pod \"goldmane-666569f655-m7pjj\" (UID: \"59842dba-2f25-4ec6-86e9-919aa2afb77a\") " pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:55:24.509345 kubelet[2982]: I0121 05:55:24.503507 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59842dba-2f25-4ec6-86e9-919aa2afb77a-goldmane-ca-bundle\") pod \"goldmane-666569f655-m7pjj\" (UID: \"59842dba-2f25-4ec6-86e9-919aa2afb77a\") " pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:55:24.509345 kubelet[2982]: I0121 05:55:24.503602 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/dac6efad-1039-4242-b1c8-782ac61e5470-calico-apiserver-certs\") pod \"calico-apiserver-796779bb-f2fds\" (UID: \"dac6efad-1039-4242-b1c8-782ac61e5470\") " pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:55:24.574014 systemd[1]: Created slice kubepods-besteffort-poddac6efad_1039_4242_b1c8_782ac61e5470.slice - libcontainer container kubepods-besteffort-poddac6efad_1039_4242_b1c8_782ac61e5470.slice. Jan 21 05:55:24.606187 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e801bb7d1500dae8ff154c994782f2902d1c801c796919eb63d5175db8a29409-rootfs.mount: Deactivated successfully. Jan 21 05:55:24.630207 systemd[1]: Created slice kubepods-besteffort-pod59842dba_2f25_4ec6_86e9_919aa2afb77a.slice - libcontainer container kubepods-besteffort-pod59842dba_2f25_4ec6_86e9_919aa2afb77a.slice. Jan 21 05:55:24.922128 systemd[1]: Created slice kubepods-burstable-pod36de057e_0cdd_42d8_a113_435fdf1d98e9.slice - libcontainer container kubepods-burstable-pod36de057e_0cdd_42d8_a113_435fdf1d98e9.slice. Jan 21 05:55:25.021579 kubelet[2982]: I0121 05:55:25.018334 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/137d495d-bc8e-4cd7-b462-b18325307853-calico-apiserver-certs\") pod \"calico-apiserver-796779bb-6bz64\" (UID: \"137d495d-bc8e-4cd7-b462-b18325307853\") " pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:55:25.021579 kubelet[2982]: I0121 05:55:25.018401 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nlrf\" (UniqueName: \"kubernetes.io/projected/137d495d-bc8e-4cd7-b462-b18325307853-kube-api-access-6nlrf\") pod \"calico-apiserver-796779bb-6bz64\" (UID: \"137d495d-bc8e-4cd7-b462-b18325307853\") " pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:55:25.186866 systemd[1]: Created slice kubepods-besteffort-poddbddb9a5_130c_465d_9677_8a7b919a18f8.slice - libcontainer container kubepods-besteffort-poddbddb9a5_130c_465d_9677_8a7b919a18f8.slice. Jan 21 05:55:25.272382 containerd[1639]: time="2026-01-21T05:55:25.272238573Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:55:25.388107 kubelet[2982]: I0121 05:55:25.388064 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5efe9c76-b287-463e-ba8c-614903640526-whisker-backend-key-pair\") pod \"whisker-db7d6575b-rxmxq\" (UID: \"5efe9c76-b287-463e-ba8c-614903640526\") " pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:55:25.388914 kubelet[2982]: I0121 05:55:25.388882 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5efe9c76-b287-463e-ba8c-614903640526-whisker-ca-bundle\") pod \"whisker-db7d6575b-rxmxq\" (UID: \"5efe9c76-b287-463e-ba8c-614903640526\") " pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:55:25.401365 containerd[1639]: time="2026-01-21T05:55:25.401236776Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:55:25.405161 kubelet[2982]: I0121 05:55:25.404874 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rrl5\" (UniqueName: \"kubernetes.io/projected/5efe9c76-b287-463e-ba8c-614903640526-kube-api-access-4rrl5\") pod \"whisker-db7d6575b-rxmxq\" (UID: \"5efe9c76-b287-463e-ba8c-614903640526\") " pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:55:25.431583 kubelet[2982]: E0121 05:55:25.431462 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:25.496769 containerd[1639]: time="2026-01-21T05:55:25.495757269Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Jan 21 05:55:25.506419 containerd[1639]: time="2026-01-21T05:55:25.504227986Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:25.513403 kubelet[2982]: I0121 05:55:25.513155 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bc2381f-9591-401f-8403-a07aa98276a8-config-volume\") pod \"coredns-674b8bbfcf-qpdqz\" (UID: \"9bc2381f-9591-401f-8403-a07aa98276a8\") " pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:55:25.513403 kubelet[2982]: I0121 05:55:25.513394 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/821a7ed1-bcf9-475e-908c-afcf62571645-tigera-ca-bundle\") pod \"calico-kube-controllers-7598d7cc9-qv4mq\" (UID: \"821a7ed1-bcf9-475e-908c-afcf62571645\") " pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:55:25.513610 kubelet[2982]: I0121 05:55:25.513502 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bll9l\" (UniqueName: \"kubernetes.io/projected/9bc2381f-9591-401f-8403-a07aa98276a8-kube-api-access-bll9l\") pod \"coredns-674b8bbfcf-qpdqz\" (UID: \"9bc2381f-9591-401f-8403-a07aa98276a8\") " pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:55:25.513610 kubelet[2982]: I0121 05:55:25.513563 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n95t\" (UniqueName: \"kubernetes.io/projected/821a7ed1-bcf9-475e-908c-afcf62571645-kube-api-access-8n95t\") pod \"calico-kube-controllers-7598d7cc9-qv4mq\" (UID: \"821a7ed1-bcf9-475e-908c-afcf62571645\") " pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:55:25.520159 systemd[1]: Created slice kubepods-besteffort-pod137d495d_bc8e_4cd7_b462_b18325307853.slice - libcontainer container kubepods-besteffort-pod137d495d_bc8e_4cd7_b462_b18325307853.slice. Jan 21 05:55:25.585869 systemd[1]: Created slice kubepods-besteffort-pod5efe9c76_b287_463e_ba8c_614903640526.slice - libcontainer container kubepods-besteffort-pod5efe9c76_b287_463e_ba8c_614903640526.slice. Jan 21 05:55:25.628098 kubelet[2982]: E0121 05:55:25.627904 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:25.701915 systemd[1]: Created slice kubepods-besteffort-pod821a7ed1_bcf9_475e_908c_afcf62571645.slice - libcontainer container kubepods-besteffort-pod821a7ed1_bcf9_475e_908c_afcf62571645.slice. Jan 21 05:55:25.722944 containerd[1639]: time="2026-01-21T05:55:25.721980295Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,}" Jan 21 05:55:25.869539 systemd[1]: Created slice kubepods-burstable-pod9bc2381f_9591_401f_8403_a07aa98276a8.slice - libcontainer container kubepods-burstable-pod9bc2381f_9591_401f_8403_a07aa98276a8.slice. Jan 21 05:55:25.940227 containerd[1639]: time="2026-01-21T05:55:25.938028671Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-db7d6575b-rxmxq,Uid:5efe9c76-b287-463e-ba8c-614903640526,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:26.107372 containerd[1639]: time="2026-01-21T05:55:26.034261628Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:55:26.185935 kubelet[2982]: E0121 05:55:26.185137 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:26.197861 containerd[1639]: time="2026-01-21T05:55:26.196534936Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:26.208076 containerd[1639]: time="2026-01-21T05:55:26.207120646Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,}" Jan 21 05:55:27.344010 containerd[1639]: time="2026-01-21T05:55:27.343827769Z" level=error msg="Failed to destroy network for sandbox \"5d678f4131306c67ee8788f33881ea1fddb8e917ba52352966d04f60aecef9d5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.355122 systemd[1]: run-netns-cni\x2d9017c84f\x2dabc5\x2da0db\x2d035c\x2daf6b253ed7a9.mount: Deactivated successfully. Jan 21 05:55:27.393792 containerd[1639]: time="2026-01-21T05:55:27.389603142Z" level=error msg="Failed to destroy network for sandbox \"94262aa85c0e17b6473c57c1994606835f32519ecd074147962119b37b5c296a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.400592 systemd[1]: run-netns-cni\x2dfdd81b0a\x2d6727\x2d8f5d\x2d4982\x2d91a6af3833a9.mount: Deactivated successfully. Jan 21 05:55:27.431622 containerd[1639]: time="2026-01-21T05:55:27.431316000Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d678f4131306c67ee8788f33881ea1fddb8e917ba52352966d04f60aecef9d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.448803 kubelet[2982]: E0121 05:55:27.434217 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d678f4131306c67ee8788f33881ea1fddb8e917ba52352966d04f60aecef9d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.448803 kubelet[2982]: E0121 05:55:27.442998 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d678f4131306c67ee8788f33881ea1fddb8e917ba52352966d04f60aecef9d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:55:27.448803 kubelet[2982]: E0121 05:55:27.447454 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5d678f4131306c67ee8788f33881ea1fddb8e917ba52352966d04f60aecef9d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:55:27.449872 containerd[1639]: time="2026-01-21T05:55:27.438632884Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"94262aa85c0e17b6473c57c1994606835f32519ecd074147962119b37b5c296a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.457033 kubelet[2982]: E0121 05:55:27.453054 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5d678f4131306c67ee8788f33881ea1fddb8e917ba52352966d04f60aecef9d5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:55:27.463838 kubelet[2982]: E0121 05:55:27.462507 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"94262aa85c0e17b6473c57c1994606835f32519ecd074147962119b37b5c296a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.463838 kubelet[2982]: E0121 05:55:27.462581 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"94262aa85c0e17b6473c57c1994606835f32519ecd074147962119b37b5c296a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:55:27.463838 kubelet[2982]: E0121 05:55:27.462609 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"94262aa85c0e17b6473c57c1994606835f32519ecd074147962119b37b5c296a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:55:27.465579 kubelet[2982]: E0121 05:55:27.464239 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"94262aa85c0e17b6473c57c1994606835f32519ecd074147962119b37b5c296a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:55:27.608141 containerd[1639]: time="2026-01-21T05:55:27.607460102Z" level=error msg="Failed to destroy network for sandbox \"bdcba88ec0ec3cf86cb56c73e5470f590ac03d8c6c1f6946249f32c79071f99d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.703297 containerd[1639]: time="2026-01-21T05:55:27.686185666Z" level=error msg="Failed to destroy network for sandbox \"eb9ec4da1c679ab2c109679ab2559775c93b9e4ab47f85b0a09afe7eeeb74aa6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.703297 containerd[1639]: time="2026-01-21T05:55:27.687540920Z" level=error msg="Failed to destroy network for sandbox \"7bb87712c86ad0e416b18323572ffb2578c81a39d26bf2f30d7aedb73435dce9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.710557 systemd[1]: run-netns-cni\x2d8da4b7a7\x2dad64\x2d205c\x2d1085\x2df009710746a0.mount: Deactivated successfully. Jan 21 05:55:27.713195 systemd[1]: run-netns-cni\x2dc684e36a\x2d7a1c\x2d8e7a\x2d5ccc\x2dc910869839ed.mount: Deactivated successfully. Jan 21 05:55:27.725199 systemd[1]: run-netns-cni\x2d78f61891\x2db79b\x2d2db2\x2d803d\x2d305e44ce7358.mount: Deactivated successfully. Jan 21 05:55:27.732164 containerd[1639]: time="2026-01-21T05:55:27.731841897Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb9ec4da1c679ab2c109679ab2559775c93b9e4ab47f85b0a09afe7eeeb74aa6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.732531 containerd[1639]: time="2026-01-21T05:55:27.732440588Z" level=error msg="Failed to destroy network for sandbox \"d8ce0fefce85966647bc55e04038002c0c0dd0292a1cadb708807ea28ee6a401\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.760343 kubelet[2982]: E0121 05:55:27.749515 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb9ec4da1c679ab2c109679ab2559775c93b9e4ab47f85b0a09afe7eeeb74aa6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.764240 containerd[1639]: time="2026-01-21T05:55:27.764052467Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bdcba88ec0ec3cf86cb56c73e5470f590ac03d8c6c1f6946249f32c79071f99d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.779331 kubelet[2982]: E0121 05:55:27.779132 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bdcba88ec0ec3cf86cb56c73e5470f590ac03d8c6c1f6946249f32c79071f99d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.779331 kubelet[2982]: E0121 05:55:27.779306 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bdcba88ec0ec3cf86cb56c73e5470f590ac03d8c6c1f6946249f32c79071f99d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zftdp" Jan 21 05:55:27.779331 kubelet[2982]: E0121 05:55:27.779341 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bdcba88ec0ec3cf86cb56c73e5470f590ac03d8c6c1f6946249f32c79071f99d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zftdp" Jan 21 05:55:27.785318 kubelet[2982]: E0121 05:55:27.782833 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bdcba88ec0ec3cf86cb56c73e5470f590ac03d8c6c1f6946249f32c79071f99d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:27.804895 kubelet[2982]: E0121 05:55:27.796841 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb9ec4da1c679ab2c109679ab2559775c93b9e4ab47f85b0a09afe7eeeb74aa6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:55:27.804895 kubelet[2982]: E0121 05:55:27.796915 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eb9ec4da1c679ab2c109679ab2559775c93b9e4ab47f85b0a09afe7eeeb74aa6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:55:27.804895 kubelet[2982]: E0121 05:55:27.797304 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"eb9ec4da1c679ab2c109679ab2559775c93b9e4ab47f85b0a09afe7eeeb74aa6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:55:27.807607 containerd[1639]: time="2026-01-21T05:55:27.802209041Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7bb87712c86ad0e416b18323572ffb2578c81a39d26bf2f30d7aedb73435dce9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.807607 containerd[1639]: time="2026-01-21T05:55:27.807565242Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d8ce0fefce85966647bc55e04038002c0c0dd0292a1cadb708807ea28ee6a401\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.808127 kubelet[2982]: E0121 05:55:27.803523 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7bb87712c86ad0e416b18323572ffb2578c81a39d26bf2f30d7aedb73435dce9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.808127 kubelet[2982]: E0121 05:55:27.803593 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7bb87712c86ad0e416b18323572ffb2578c81a39d26bf2f30d7aedb73435dce9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:55:27.808127 kubelet[2982]: E0121 05:55:27.803625 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7bb87712c86ad0e416b18323572ffb2578c81a39d26bf2f30d7aedb73435dce9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:55:27.809766 kubelet[2982]: E0121 05:55:27.808517 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-qpdqz_kube-system(9bc2381f-9591-401f-8403-a07aa98276a8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-qpdqz_kube-system(9bc2381f-9591-401f-8403-a07aa98276a8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7bb87712c86ad0e416b18323572ffb2578c81a39d26bf2f30d7aedb73435dce9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-qpdqz" podUID="9bc2381f-9591-401f-8403-a07aa98276a8" Jan 21 05:55:27.809766 kubelet[2982]: E0121 05:55:27.809022 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d8ce0fefce85966647bc55e04038002c0c0dd0292a1cadb708807ea28ee6a401\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.809766 kubelet[2982]: E0121 05:55:27.809078 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d8ce0fefce85966647bc55e04038002c0c0dd0292a1cadb708807ea28ee6a401\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:55:27.810157 kubelet[2982]: E0121 05:55:27.809108 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d8ce0fefce85966647bc55e04038002c0c0dd0292a1cadb708807ea28ee6a401\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:55:27.812123 kubelet[2982]: E0121 05:55:27.810865 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d8ce0fefce85966647bc55e04038002c0c0dd0292a1cadb708807ea28ee6a401\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:55:27.864611 containerd[1639]: time="2026-01-21T05:55:27.858315947Z" level=error msg="Failed to destroy network for sandbox \"90704de8503e347a5073fb81c2aa48f7fbd2119a5adf012a3b90baad9a3277f0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.867808 containerd[1639]: time="2026-01-21T05:55:27.867615635Z" level=error msg="Failed to destroy network for sandbox \"bd29bf60f8c9736094430bd192ac214d118b9bd937aa377631f6643cdf98657a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.902900 containerd[1639]: time="2026-01-21T05:55:27.902817025Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"90704de8503e347a5073fb81c2aa48f7fbd2119a5adf012a3b90baad9a3277f0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.906270 kubelet[2982]: E0121 05:55:27.905116 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"90704de8503e347a5073fb81c2aa48f7fbd2119a5adf012a3b90baad9a3277f0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.906270 kubelet[2982]: E0121 05:55:27.905540 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"90704de8503e347a5073fb81c2aa48f7fbd2119a5adf012a3b90baad9a3277f0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:55:27.908487 kubelet[2982]: E0121 05:55:27.907057 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"90704de8503e347a5073fb81c2aa48f7fbd2119a5adf012a3b90baad9a3277f0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:55:27.908487 kubelet[2982]: E0121 05:55:27.907140 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-sgzh7_kube-system(36de057e-0cdd-42d8-a113-435fdf1d98e9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-sgzh7_kube-system(36de057e-0cdd-42d8-a113-435fdf1d98e9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"90704de8503e347a5073fb81c2aa48f7fbd2119a5adf012a3b90baad9a3277f0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-sgzh7" podUID="36de057e-0cdd-42d8-a113-435fdf1d98e9" Jan 21 05:55:27.917811 containerd[1639]: time="2026-01-21T05:55:27.913869665Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bd29bf60f8c9736094430bd192ac214d118b9bd937aa377631f6643cdf98657a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.918548 kubelet[2982]: E0121 05:55:27.915951 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bd29bf60f8c9736094430bd192ac214d118b9bd937aa377631f6643cdf98657a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:27.918548 kubelet[2982]: E0121 05:55:27.916100 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bd29bf60f8c9736094430bd192ac214d118b9bd937aa377631f6643cdf98657a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:55:27.918548 kubelet[2982]: E0121 05:55:27.916139 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bd29bf60f8c9736094430bd192ac214d118b9bd937aa377631f6643cdf98657a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:55:27.918843 kubelet[2982]: E0121 05:55:27.916208 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bd29bf60f8c9736094430bd192ac214d118b9bd937aa377631f6643cdf98657a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:55:27.969035 systemd[1]: run-netns-cni\x2d2d93f66c\x2dd2c8\x2dfb33\x2d5c8a\x2dc193438bfb37.mount: Deactivated successfully. Jan 21 05:55:27.972511 systemd[1]: run-netns-cni\x2dbbca54c1\x2d13f4\x2dd6be\x2d0bf7\x2d7fffac2aeff1.mount: Deactivated successfully. Jan 21 05:55:27.972776 systemd[1]: run-netns-cni\x2d564c54ef\x2d5cb6\x2d3391\x2dbbd9\x2dd3ba123de9ab.mount: Deactivated successfully. Jan 21 05:55:28.017168 containerd[1639]: time="2026-01-21T05:55:28.013051090Z" level=error msg="Failed to destroy network for sandbox \"8adeed02130966b64c9af425d6332a669aaf923e6a1f5566f85b6c1f6bfd80ba\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:28.029808 systemd[1]: run-netns-cni\x2dd4fee224\x2def3c\x2daf9c\x2d6ed3\x2dd32280700bc9.mount: Deactivated successfully. Jan 21 05:55:28.059808 containerd[1639]: time="2026-01-21T05:55:28.057366119Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-db7d6575b-rxmxq,Uid:5efe9c76-b287-463e-ba8c-614903640526,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8adeed02130966b64c9af425d6332a669aaf923e6a1f5566f85b6c1f6bfd80ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:28.063028 kubelet[2982]: E0121 05:55:28.062838 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8adeed02130966b64c9af425d6332a669aaf923e6a1f5566f85b6c1f6bfd80ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:28.063193 kubelet[2982]: E0121 05:55:28.063024 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8adeed02130966b64c9af425d6332a669aaf923e6a1f5566f85b6c1f6bfd80ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:55:28.063193 kubelet[2982]: E0121 05:55:28.063069 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8adeed02130966b64c9af425d6332a669aaf923e6a1f5566f85b6c1f6bfd80ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:55:28.065585 kubelet[2982]: E0121 05:55:28.065114 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-db7d6575b-rxmxq_calico-system(5efe9c76-b287-463e-ba8c-614903640526)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-db7d6575b-rxmxq_calico-system(5efe9c76-b287-463e-ba8c-614903640526)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8adeed02130966b64c9af425d6332a669aaf923e6a1f5566f85b6c1f6bfd80ba\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-db7d6575b-rxmxq" podUID="5efe9c76-b287-463e-ba8c-614903640526" Jan 21 05:55:37.210299 kubelet[2982]: E0121 05:55:37.210241 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:38.158000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.61:22-10.0.0.1:53176 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:55:38.158626 systemd[1]: Started sshd@9-10.0.0.61:22-10.0.0.1:53176.service - OpenSSH per-connection server daemon (10.0.0.1:53176). Jan 21 05:55:38.202948 kernel: audit: type=1130 audit(1768974938.158:575): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.61:22-10.0.0.1:53176 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:55:38.255297 containerd[1639]: time="2026-01-21T05:55:38.250496898Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:38.255297 containerd[1639]: time="2026-01-21T05:55:38.255078703Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:38.747000 audit[4228]: USER_ACCT pid=4228 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:38.772157 sshd-session[4228]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:55:38.796490 kernel: audit: type=1101 audit(1768974938.747:576): pid=4228 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:38.796546 kernel: audit: type=1103 audit(1768974938.763:577): pid=4228 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:38.763000 audit[4228]: CRED_ACQ pid=4228 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:38.796783 sshd[4228]: Accepted publickey for core from 10.0.0.1 port 53176 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:55:38.859162 kernel: audit: type=1006 audit(1768974938.767:578): pid=4228 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=11 res=1 Jan 21 05:55:38.859290 kernel: audit: type=1300 audit(1768974938.767:578): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd6d145db0 a2=3 a3=0 items=0 ppid=1 pid=4228 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:38.767000 audit[4228]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd6d145db0 a2=3 a3=0 items=0 ppid=1 pid=4228 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:38.867479 systemd-logind[1620]: New session 11 of user core. Jan 21 05:55:38.767000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:55:38.933818 kernel: audit: type=1327 audit(1768974938.767:578): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:55:38.947265 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 21 05:55:38.980000 audit[4228]: USER_START pid=4228 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:38.994000 audit[4262]: CRED_ACQ pid=4262 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:39.052255 kernel: audit: type=1105 audit(1768974938.980:579): pid=4228 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:39.052477 kernel: audit: type=1103 audit(1768974938.994:580): pid=4262 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:39.218482 containerd[1639]: time="2026-01-21T05:55:39.218289359Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:55:39.229285 containerd[1639]: time="2026-01-21T05:55:39.229046056Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:55:39.497538 containerd[1639]: time="2026-01-21T05:55:39.497377372Z" level=error msg="Failed to destroy network for sandbox \"329a9a1683361e60c2d4f091aecb20cebc890a44f7ddb51e7a1bec5a30219ad3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:39.509161 containerd[1639]: time="2026-01-21T05:55:39.508087450Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"329a9a1683361e60c2d4f091aecb20cebc890a44f7ddb51e7a1bec5a30219ad3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:39.510081 kubelet[2982]: E0121 05:55:39.509408 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"329a9a1683361e60c2d4f091aecb20cebc890a44f7ddb51e7a1bec5a30219ad3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:39.510081 kubelet[2982]: E0121 05:55:39.509487 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"329a9a1683361e60c2d4f091aecb20cebc890a44f7ddb51e7a1bec5a30219ad3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:55:39.510081 kubelet[2982]: E0121 05:55:39.509525 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"329a9a1683361e60c2d4f091aecb20cebc890a44f7ddb51e7a1bec5a30219ad3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:55:39.513985 kubelet[2982]: E0121 05:55:39.509595 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"329a9a1683361e60c2d4f091aecb20cebc890a44f7ddb51e7a1bec5a30219ad3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:55:39.516369 systemd[1]: run-netns-cni\x2d0c8355b2\x2d21c0\x2d9ea1\x2d0e1d\x2d078d72402893.mount: Deactivated successfully. Jan 21 05:55:39.711057 containerd[1639]: time="2026-01-21T05:55:39.709516830Z" level=error msg="Failed to destroy network for sandbox \"2a213c9695fe7e705cd876e1f9e87ca4f9518713b14ddd950980232e8541513e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:39.726423 systemd[1]: run-netns-cni\x2d57c6ba38\x2dbb62\x2d5acb\x2dd9b3\x2d699e5c326294.mount: Deactivated successfully. Jan 21 05:55:39.772090 containerd[1639]: time="2026-01-21T05:55:39.762114545Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a213c9695fe7e705cd876e1f9e87ca4f9518713b14ddd950980232e8541513e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:39.775116 kubelet[2982]: E0121 05:55:39.766263 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a213c9695fe7e705cd876e1f9e87ca4f9518713b14ddd950980232e8541513e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:39.775116 kubelet[2982]: E0121 05:55:39.768023 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a213c9695fe7e705cd876e1f9e87ca4f9518713b14ddd950980232e8541513e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zftdp" Jan 21 05:55:39.775116 kubelet[2982]: E0121 05:55:39.768064 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a213c9695fe7e705cd876e1f9e87ca4f9518713b14ddd950980232e8541513e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zftdp" Jan 21 05:55:39.780250 kubelet[2982]: E0121 05:55:39.774844 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2a213c9695fe7e705cd876e1f9e87ca4f9518713b14ddd950980232e8541513e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:55:39.995453 sshd[4262]: Connection closed by 10.0.0.1 port 53176 Jan 21 05:55:39.999351 sshd-session[4228]: pam_unix(sshd:session): session closed for user core Jan 21 05:55:40.021000 audit[4228]: USER_END pid=4228 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:40.050611 systemd[1]: sshd@9-10.0.0.61:22-10.0.0.1:53176.service: Deactivated successfully. Jan 21 05:55:40.059838 kernel: audit: type=1106 audit(1768974940.021:581): pid=4228 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:40.068882 systemd[1]: session-11.scope: Deactivated successfully. Jan 21 05:55:40.022000 audit[4228]: CRED_DISP pid=4228 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:40.084166 systemd-logind[1620]: Session 11 logged out. Waiting for processes to exit. Jan 21 05:55:40.089069 systemd-logind[1620]: Removed session 11. Jan 21 05:55:40.097291 kernel: audit: type=1104 audit(1768974940.022:582): pid=4228 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:40.050000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.61:22-10.0.0.1:53176 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:55:40.187844 containerd[1639]: time="2026-01-21T05:55:40.187273649Z" level=error msg="Failed to destroy network for sandbox \"ca7a60b61b024a401ec37ec9d732712a101ec8408c564f862347b93c5ba19e3a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.208516 systemd[1]: run-netns-cni\x2d74cacab3\x2d8cbb\x2d4375\x2dc282\x2d1e1fcbde98da.mount: Deactivated successfully. Jan 21 05:55:40.220017 containerd[1639]: time="2026-01-21T05:55:40.219047805Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca7a60b61b024a401ec37ec9d732712a101ec8408c564f862347b93c5ba19e3a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.220017 containerd[1639]: time="2026-01-21T05:55:40.219133589Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:40.220320 kubelet[2982]: E0121 05:55:40.220025 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca7a60b61b024a401ec37ec9d732712a101ec8408c564f862347b93c5ba19e3a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.220320 kubelet[2982]: E0121 05:55:40.220094 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca7a60b61b024a401ec37ec9d732712a101ec8408c564f862347b93c5ba19e3a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:55:40.220320 kubelet[2982]: E0121 05:55:40.220119 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca7a60b61b024a401ec37ec9d732712a101ec8408c564f862347b93c5ba19e3a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:55:40.220470 kubelet[2982]: E0121 05:55:40.220183 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ca7a60b61b024a401ec37ec9d732712a101ec8408c564f862347b93c5ba19e3a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:55:40.225817 kubelet[2982]: E0121 05:55:40.224419 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:40.235546 containerd[1639]: time="2026-01-21T05:55:40.232152051Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,}" Jan 21 05:55:40.277005 containerd[1639]: time="2026-01-21T05:55:40.257904594Z" level=error msg="Failed to destroy network for sandbox \"4ba6f56a257cbf7684997cfcd0583eabe11b614237ab7ce1424de8cc5ad1d55a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.294489 containerd[1639]: time="2026-01-21T05:55:40.294048641Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ba6f56a257cbf7684997cfcd0583eabe11b614237ab7ce1424de8cc5ad1d55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.310021 kubelet[2982]: E0121 05:55:40.307125 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ba6f56a257cbf7684997cfcd0583eabe11b614237ab7ce1424de8cc5ad1d55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.310021 kubelet[2982]: E0121 05:55:40.307289 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ba6f56a257cbf7684997cfcd0583eabe11b614237ab7ce1424de8cc5ad1d55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:55:40.310021 kubelet[2982]: E0121 05:55:40.307330 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ba6f56a257cbf7684997cfcd0583eabe11b614237ab7ce1424de8cc5ad1d55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:55:40.310250 kubelet[2982]: E0121 05:55:40.307399 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4ba6f56a257cbf7684997cfcd0583eabe11b614237ab7ce1424de8cc5ad1d55a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:55:40.422372 systemd[1]: run-netns-cni\x2dcf94d419\x2d867d\x2d5f4a\x2de138\x2dfed0d7c22d63.mount: Deactivated successfully. Jan 21 05:55:40.651137 containerd[1639]: time="2026-01-21T05:55:40.649903958Z" level=error msg="Failed to destroy network for sandbox \"70bcc488e3cbb955b4c620da843223fde6482a5ec28fdd3cb95302dae74ff088\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.660355 systemd[1]: run-netns-cni\x2d8b362395\x2df804\x2d6a9d\x2db9c8\x2d41544b1f3533.mount: Deactivated successfully. Jan 21 05:55:40.678217 containerd[1639]: time="2026-01-21T05:55:40.677805311Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"70bcc488e3cbb955b4c620da843223fde6482a5ec28fdd3cb95302dae74ff088\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.678605 kubelet[2982]: E0121 05:55:40.678520 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"70bcc488e3cbb955b4c620da843223fde6482a5ec28fdd3cb95302dae74ff088\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.678605 kubelet[2982]: E0121 05:55:40.678619 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"70bcc488e3cbb955b4c620da843223fde6482a5ec28fdd3cb95302dae74ff088\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:55:40.678605 kubelet[2982]: E0121 05:55:40.678788 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"70bcc488e3cbb955b4c620da843223fde6482a5ec28fdd3cb95302dae74ff088\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:55:40.685106 kubelet[2982]: E0121 05:55:40.679237 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"70bcc488e3cbb955b4c620da843223fde6482a5ec28fdd3cb95302dae74ff088\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:55:40.704136 containerd[1639]: time="2026-01-21T05:55:40.701042160Z" level=error msg="Failed to destroy network for sandbox \"f5f6e4423f28ee313e7d6f692c1cdc91c47973a49eb203c3736266a3fe64140e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.706605 systemd[1]: run-netns-cni\x2d79f2bee2\x2d5d6d\x2d2f89\x2de274\x2d30b4a989ba0c.mount: Deactivated successfully. Jan 21 05:55:40.719833 containerd[1639]: time="2026-01-21T05:55:40.718340450Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f5f6e4423f28ee313e7d6f692c1cdc91c47973a49eb203c3736266a3fe64140e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.725331 kubelet[2982]: E0121 05:55:40.723148 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f5f6e4423f28ee313e7d6f692c1cdc91c47973a49eb203c3736266a3fe64140e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:40.725331 kubelet[2982]: E0121 05:55:40.725126 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f5f6e4423f28ee313e7d6f692c1cdc91c47973a49eb203c3736266a3fe64140e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:55:40.725331 kubelet[2982]: E0121 05:55:40.725170 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f5f6e4423f28ee313e7d6f692c1cdc91c47973a49eb203c3736266a3fe64140e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:55:40.726300 kubelet[2982]: E0121 05:55:40.726076 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-qpdqz_kube-system(9bc2381f-9591-401f-8403-a07aa98276a8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-qpdqz_kube-system(9bc2381f-9591-401f-8403-a07aa98276a8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f5f6e4423f28ee313e7d6f692c1cdc91c47973a49eb203c3736266a3fe64140e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-qpdqz" podUID="9bc2381f-9591-401f-8403-a07aa98276a8" Jan 21 05:55:42.216815 containerd[1639]: time="2026-01-21T05:55:42.212429786Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-db7d6575b-rxmxq,Uid:5efe9c76-b287-463e-ba8c-614903640526,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:42.216815 containerd[1639]: time="2026-01-21T05:55:42.214817470Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:55:43.078604 containerd[1639]: time="2026-01-21T05:55:43.078325927Z" level=error msg="Failed to destroy network for sandbox \"9862fd53c218319f7d48ef16963775cd7967a8f6c2816caf776ce7748ec20a5c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:43.129954 systemd[1]: run-netns-cni\x2d2e9b111a\x2d172f\x2d7a58\x2d57ce\x2d41570333adf8.mount: Deactivated successfully. Jan 21 05:55:43.171188 containerd[1639]: time="2026-01-21T05:55:43.170997349Z" level=error msg="Failed to destroy network for sandbox \"ed66bbd8efa8c6b85a8fadf422339de45d406de45129176346f1531e821b2b7e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:43.225888 kubelet[2982]: E0121 05:55:43.225482 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:43.288202 containerd[1639]: time="2026-01-21T05:55:43.288150626Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,}" Jan 21 05:55:43.305984 systemd[1]: run-netns-cni\x2d70aa9ca5\x2def26\x2d4376\x2d1da2\x2d844eda79075e.mount: Deactivated successfully. Jan 21 05:55:43.335208 containerd[1639]: time="2026-01-21T05:55:43.330143386Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ed66bbd8efa8c6b85a8fadf422339de45d406de45129176346f1531e821b2b7e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:43.349279 containerd[1639]: time="2026-01-21T05:55:43.344132839Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-db7d6575b-rxmxq,Uid:5efe9c76-b287-463e-ba8c-614903640526,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9862fd53c218319f7d48ef16963775cd7967a8f6c2816caf776ce7748ec20a5c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:43.362787 kubelet[2982]: E0121 05:55:43.339491 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ed66bbd8efa8c6b85a8fadf422339de45d406de45129176346f1531e821b2b7e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:43.482417 kubelet[2982]: E0121 05:55:43.472629 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9862fd53c218319f7d48ef16963775cd7967a8f6c2816caf776ce7748ec20a5c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:43.493373 kubelet[2982]: E0121 05:55:43.492836 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9862fd53c218319f7d48ef16963775cd7967a8f6c2816caf776ce7748ec20a5c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:55:43.493373 kubelet[2982]: E0121 05:55:43.493161 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9862fd53c218319f7d48ef16963775cd7967a8f6c2816caf776ce7748ec20a5c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:55:43.494891 kubelet[2982]: E0121 05:55:43.493330 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-db7d6575b-rxmxq_calico-system(5efe9c76-b287-463e-ba8c-614903640526)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-db7d6575b-rxmxq_calico-system(5efe9c76-b287-463e-ba8c-614903640526)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9862fd53c218319f7d48ef16963775cd7967a8f6c2816caf776ce7748ec20a5c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-db7d6575b-rxmxq" podUID="5efe9c76-b287-463e-ba8c-614903640526" Jan 21 05:55:43.505310 kubelet[2982]: E0121 05:55:43.505263 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ed66bbd8efa8c6b85a8fadf422339de45d406de45129176346f1531e821b2b7e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:55:43.516818 kubelet[2982]: E0121 05:55:43.516771 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ed66bbd8efa8c6b85a8fadf422339de45d406de45129176346f1531e821b2b7e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:55:43.517228 kubelet[2982]: E0121 05:55:43.517185 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ed66bbd8efa8c6b85a8fadf422339de45d406de45129176346f1531e821b2b7e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:55:44.900861 containerd[1639]: time="2026-01-21T05:55:44.900801364Z" level=error msg="Failed to destroy network for sandbox \"288bb066333ec531d13ae7600fb5f58d5584faed119c9e2e9be29c978bad512f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:44.910984 systemd[1]: run-netns-cni\x2d90be9d52\x2d1311\x2dab2d\x2def1d\x2d40aad5b7f3cd.mount: Deactivated successfully. Jan 21 05:55:44.926393 containerd[1639]: time="2026-01-21T05:55:44.926242373Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"288bb066333ec531d13ae7600fb5f58d5584faed119c9e2e9be29c978bad512f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:44.929931 kubelet[2982]: E0121 05:55:44.929878 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"288bb066333ec531d13ae7600fb5f58d5584faed119c9e2e9be29c978bad512f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:55:44.931618 kubelet[2982]: E0121 05:55:44.931314 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"288bb066333ec531d13ae7600fb5f58d5584faed119c9e2e9be29c978bad512f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:55:44.931618 kubelet[2982]: E0121 05:55:44.931418 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"288bb066333ec531d13ae7600fb5f58d5584faed119c9e2e9be29c978bad512f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:55:44.931618 kubelet[2982]: E0121 05:55:44.931492 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-sgzh7_kube-system(36de057e-0cdd-42d8-a113-435fdf1d98e9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-sgzh7_kube-system(36de057e-0cdd-42d8-a113-435fdf1d98e9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"288bb066333ec531d13ae7600fb5f58d5584faed119c9e2e9be29c978bad512f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-sgzh7" podUID="36de057e-0cdd-42d8-a113-435fdf1d98e9" Jan 21 05:55:45.046775 systemd[1]: Started sshd@10-10.0.0.61:22-10.0.0.1:43534.service - OpenSSH per-connection server daemon (10.0.0.1:43534). Jan 21 05:55:45.096485 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:55:45.096628 kernel: audit: type=1130 audit(1768974945.045:584): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.61:22-10.0.0.1:43534 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:55:45.045000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.61:22-10.0.0.1:43534 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:55:45.612000 audit[4541]: USER_ACCT pid=4541 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:45.628236 sshd-session[4541]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:55:45.634375 sshd[4541]: Accepted publickey for core from 10.0.0.1 port 43534 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:55:45.616000 audit[4541]: CRED_ACQ pid=4541 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:45.659343 systemd-logind[1620]: New session 12 of user core. Jan 21 05:55:45.670365 kernel: audit: type=1101 audit(1768974945.612:585): pid=4541 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:45.670473 kernel: audit: type=1103 audit(1768974945.616:586): pid=4541 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:45.670597 kernel: audit: type=1006 audit(1768974945.620:587): pid=4541 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=12 res=1 Jan 21 05:55:45.620000 audit[4541]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd12d9cbc0 a2=3 a3=0 items=0 ppid=1 pid=4541 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:45.715435 kernel: audit: type=1300 audit(1768974945.620:587): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd12d9cbc0 a2=3 a3=0 items=0 ppid=1 pid=4541 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:45.715619 kernel: audit: type=1327 audit(1768974945.620:587): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:55:45.620000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:55:45.739470 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 21 05:55:45.760000 audit[4541]: USER_START pid=4541 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:45.822039 kernel: audit: type=1105 audit(1768974945.760:588): pid=4541 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:45.824568 kernel: audit: type=1103 audit(1768974945.769:589): pid=4545 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:45.769000 audit[4545]: CRED_ACQ pid=4545 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:47.121010 sshd[4545]: Connection closed by 10.0.0.1 port 43534 Jan 21 05:55:47.179968 sshd-session[4541]: pam_unix(sshd:session): session closed for user core Jan 21 05:55:47.237000 audit[4541]: USER_END pid=4541 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:47.268913 systemd-logind[1620]: Session 12 logged out. Waiting for processes to exit. Jan 21 05:55:47.307169 kernel: audit: type=1106 audit(1768974947.237:590): pid=4541 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:47.307764 kernel: audit: type=1104 audit(1768974947.239:591): pid=4541 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:47.239000 audit[4541]: CRED_DISP pid=4541 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:47.269954 systemd[1]: sshd@10-10.0.0.61:22-10.0.0.1:43534.service: Deactivated successfully. Jan 21 05:55:47.305599 systemd[1]: session-12.scope: Deactivated successfully. Jan 21 05:55:47.325112 systemd-logind[1620]: Removed session 12. Jan 21 05:55:47.279000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.61:22-10.0.0.1:43534 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:55:57.873349 systemd[1]: Started sshd@11-10.0.0.61:22-10.0.0.1:43536.service - OpenSSH per-connection server daemon (10.0.0.1:43536). Jan 21 05:55:57.972706 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:55:57.972969 kernel: audit: type=1130 audit(1768974957.918:593): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.61:22-10.0.0.1:43536 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:55:57.918000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.61:22-10.0.0.1:43536 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:55:58.308827 containerd[1639]: time="2026-01-21T05:55:58.303622540Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:58.458009 systemd[1740]: Created slice background.slice - User Background Tasks Slice. Jan 21 05:55:58.466041 systemd[1740]: Starting systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories... Jan 21 05:55:58.598978 kubelet[2982]: E0121 05:55:58.512211 2982 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="7.025s" Jan 21 05:55:58.751760 containerd[1639]: time="2026-01-21T05:55:58.748282924Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:58.855794 containerd[1639]: time="2026-01-21T05:55:58.849203339Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:55:58.930197 systemd[1740]: Finished systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories. Jan 21 05:55:58.997031 kubelet[2982]: E0121 05:55:58.995970 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:58.997219 containerd[1639]: time="2026-01-21T05:55:58.996012822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:58.998249 containerd[1639]: time="2026-01-21T05:55:58.998219063Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:55:59.009991 containerd[1639]: time="2026-01-21T05:55:58.999334145Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,}" Jan 21 05:55:59.010128 kubelet[2982]: E0121 05:55:59.008436 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:59.016407 containerd[1639]: time="2026-01-21T05:55:59.016370449Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:55:59.026217 kubelet[2982]: E0121 05:55:59.019229 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:59.048465 containerd[1639]: time="2026-01-21T05:55:59.048407056Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,}" Jan 21 05:55:59.203548 containerd[1639]: time="2026-01-21T05:55:59.197077003Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-db7d6575b-rxmxq,Uid:5efe9c76-b287-463e-ba8c-614903640526,Namespace:calico-system,Attempt:0,}" Jan 21 05:55:59.232811 kubelet[2982]: E0121 05:55:59.232369 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:55:59.261000 audit[4561]: USER_ACCT pid=4561 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:59.295190 sshd[4561]: Accepted publickey for core from 10.0.0.1 port 43536 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:55:59.308626 sshd-session[4561]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:55:59.320923 kernel: audit: type=1101 audit(1768974959.261:594): pid=4561 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:59.321023 kernel: audit: type=1103 audit(1768974959.298:595): pid=4561 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:59.298000 audit[4561]: CRED_ACQ pid=4561 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:59.422848 kernel: audit: type=1006 audit(1768974959.298:596): pid=4561 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=13 res=1 Jan 21 05:55:59.298000 audit[4561]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff1a7ff6b0 a2=3 a3=0 items=0 ppid=1 pid=4561 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:59.431462 systemd-logind[1620]: New session 13 of user core. Jan 21 05:55:59.473399 kernel: audit: type=1300 audit(1768974959.298:596): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff1a7ff6b0 a2=3 a3=0 items=0 ppid=1 pid=4561 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:55:59.488802 kernel: audit: type=1327 audit(1768974959.298:596): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:55:59.298000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:55:59.496504 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 21 05:55:59.588000 audit[4561]: USER_START pid=4561 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:59.668433 kernel: audit: type=1105 audit(1768974959.588:597): pid=4561 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:59.679402 kernel: audit: type=1103 audit(1768974959.613:598): pid=4597 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:55:59.613000 audit[4597]: CRED_ACQ pid=4597 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:02.401415 kubelet[2982]: E0121 05:56:02.401311 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:05.956086 sshd[4597]: Connection closed by 10.0.0.1 port 43536 Jan 21 05:56:05.963198 sshd-session[4561]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:06.064000 audit[4561]: USER_END pid=4561 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:06.126418 kernel: audit: type=1106 audit(1768974966.064:599): pid=4561 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:06.066000 audit[4561]: CRED_DISP pid=4561 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:06.148830 systemd[1]: sshd@11-10.0.0.61:22-10.0.0.1:43536.service: Deactivated successfully. Jan 21 05:56:06.181447 kernel: audit: type=1104 audit(1768974966.066:600): pid=4561 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:06.146000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.61:22-10.0.0.1:43536 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:06.164508 systemd[1]: session-13.scope: Deactivated successfully. Jan 21 05:56:06.170081 systemd-logind[1620]: Session 13 logged out. Waiting for processes to exit. Jan 21 05:56:06.189050 systemd-logind[1620]: Removed session 13. Jan 21 05:56:06.238111 kernel: audit: type=1131 audit(1768974966.146:601): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.61:22-10.0.0.1:43536 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:07.324931 containerd[1639]: time="2026-01-21T05:56:07.322625934Z" level=error msg="Failed to destroy network for sandbox \"a1d20cf7c669c7b92948daca863d85b4e5222b63e1e1c203677765026275c0a7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.419586 containerd[1639]: time="2026-01-21T05:56:07.418460666Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-db7d6575b-rxmxq,Uid:5efe9c76-b287-463e-ba8c-614903640526,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a1d20cf7c669c7b92948daca863d85b4e5222b63e1e1c203677765026275c0a7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.404313 systemd[1]: run-netns-cni\x2d580cc9cd\x2dec25\x2d7873\x2dee0a\x2d0cd158f12ef3.mount: Deactivated successfully. Jan 21 05:56:07.514534 kubelet[2982]: E0121 05:56:07.425101 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a1d20cf7c669c7b92948daca863d85b4e5222b63e1e1c203677765026275c0a7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.514534 kubelet[2982]: E0121 05:56:07.431383 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a1d20cf7c669c7b92948daca863d85b4e5222b63e1e1c203677765026275c0a7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:56:07.514534 kubelet[2982]: E0121 05:56:07.431980 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a1d20cf7c669c7b92948daca863d85b4e5222b63e1e1c203677765026275c0a7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:56:07.516098 containerd[1639]: time="2026-01-21T05:56:07.501350435Z" level=error msg="Failed to destroy network for sandbox \"213d8594a8939e85889de8b2e5ae99984c9bd85bf666e5f0f0afc90c21d76a3d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.516098 containerd[1639]: time="2026-01-21T05:56:07.515946282Z" level=error msg="Failed to destroy network for sandbox \"b602dfb14347e96edfbbeb5761100321b43af7c1c33467bcf09c3a39b216afae\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.516487 kubelet[2982]: E0121 05:56:07.433343 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-db7d6575b-rxmxq_calico-system(5efe9c76-b287-463e-ba8c-614903640526)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-db7d6575b-rxmxq_calico-system(5efe9c76-b287-463e-ba8c-614903640526)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a1d20cf7c669c7b92948daca863d85b4e5222b63e1e1c203677765026275c0a7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-db7d6575b-rxmxq" podUID="5efe9c76-b287-463e-ba8c-614903640526" Jan 21 05:56:07.522392 systemd[1]: run-netns-cni\x2d33d8652c\x2d5ad4\x2d18fd\x2d0982\x2dcff98af57e3b.mount: Deactivated successfully. Jan 21 05:56:07.546922 systemd[1]: run-netns-cni\x2d001624a0\x2d5607\x2dbb56\x2dc1cd\x2dd972c453356c.mount: Deactivated successfully. Jan 21 05:56:07.556545 containerd[1639]: time="2026-01-21T05:56:07.556199501Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b602dfb14347e96edfbbeb5761100321b43af7c1c33467bcf09c3a39b216afae\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.568629 kubelet[2982]: E0121 05:56:07.563357 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b602dfb14347e96edfbbeb5761100321b43af7c1c33467bcf09c3a39b216afae\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.568629 kubelet[2982]: E0121 05:56:07.563444 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b602dfb14347e96edfbbeb5761100321b43af7c1c33467bcf09c3a39b216afae\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:56:07.568629 kubelet[2982]: E0121 05:56:07.563476 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b602dfb14347e96edfbbeb5761100321b43af7c1c33467bcf09c3a39b216afae\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:56:07.581325 kubelet[2982]: E0121 05:56:07.563616 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b602dfb14347e96edfbbeb5761100321b43af7c1c33467bcf09c3a39b216afae\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:56:07.597619 containerd[1639]: time="2026-01-21T05:56:07.597076048Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"213d8594a8939e85889de8b2e5ae99984c9bd85bf666e5f0f0afc90c21d76a3d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.608765 kubelet[2982]: E0121 05:56:07.607269 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"213d8594a8939e85889de8b2e5ae99984c9bd85bf666e5f0f0afc90c21d76a3d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.608765 kubelet[2982]: E0121 05:56:07.608029 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"213d8594a8939e85889de8b2e5ae99984c9bd85bf666e5f0f0afc90c21d76a3d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:56:07.608765 kubelet[2982]: E0121 05:56:07.608401 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"213d8594a8939e85889de8b2e5ae99984c9bd85bf666e5f0f0afc90c21d76a3d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:56:07.620275 kubelet[2982]: E0121 05:56:07.620219 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"213d8594a8939e85889de8b2e5ae99984c9bd85bf666e5f0f0afc90c21d76a3d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:56:07.706488 containerd[1639]: time="2026-01-21T05:56:07.703317776Z" level=error msg="Failed to destroy network for sandbox \"be5e845d29fb32114c84afbded005314ad86e62df4902c1c653563f630c56d72\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.706488 containerd[1639]: time="2026-01-21T05:56:07.703402457Z" level=error msg="Failed to destroy network for sandbox \"972fc01390efd8be9ca0883aa6a15fe889e373cabd9a13a5fec7fbcd1eefdd50\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.714943 systemd[1]: run-netns-cni\x2df8a6b158\x2d3a50\x2de449\x2d5373\x2d6e6eeb873efc.mount: Deactivated successfully. Jan 21 05:56:07.732489 systemd[1]: run-netns-cni\x2d2b2cbdfd\x2db7ee\x2dd8d7\x2d41de\x2d2170f3f411b0.mount: Deactivated successfully. Jan 21 05:56:07.772496 containerd[1639]: time="2026-01-21T05:56:07.772431017Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"be5e845d29fb32114c84afbded005314ad86e62df4902c1c653563f630c56d72\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.780858 kubelet[2982]: E0121 05:56:07.780100 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"be5e845d29fb32114c84afbded005314ad86e62df4902c1c653563f630c56d72\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.780858 kubelet[2982]: E0121 05:56:07.780256 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"be5e845d29fb32114c84afbded005314ad86e62df4902c1c653563f630c56d72\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:56:07.780858 kubelet[2982]: E0121 05:56:07.780298 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"be5e845d29fb32114c84afbded005314ad86e62df4902c1c653563f630c56d72\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:56:07.781337 kubelet[2982]: E0121 05:56:07.780367 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"be5e845d29fb32114c84afbded005314ad86e62df4902c1c653563f630c56d72\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:56:07.783256 containerd[1639]: time="2026-01-21T05:56:07.783201925Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"972fc01390efd8be9ca0883aa6a15fe889e373cabd9a13a5fec7fbcd1eefdd50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.792404 kubelet[2982]: E0121 05:56:07.789545 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"972fc01390efd8be9ca0883aa6a15fe889e373cabd9a13a5fec7fbcd1eefdd50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.797234 kubelet[2982]: E0121 05:56:07.795625 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"972fc01390efd8be9ca0883aa6a15fe889e373cabd9a13a5fec7fbcd1eefdd50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:56:07.797234 kubelet[2982]: E0121 05:56:07.796608 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"972fc01390efd8be9ca0883aa6a15fe889e373cabd9a13a5fec7fbcd1eefdd50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:56:07.803252 kubelet[2982]: E0121 05:56:07.800965 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"972fc01390efd8be9ca0883aa6a15fe889e373cabd9a13a5fec7fbcd1eefdd50\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:56:07.855054 containerd[1639]: time="2026-01-21T05:56:07.851972441Z" level=error msg="Failed to destroy network for sandbox \"0da86421731ab4d4495d83e68c9c2c265facde362c89f6502ff713132f1484ad\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.900040 containerd[1639]: time="2026-01-21T05:56:07.898472762Z" level=error msg="Failed to destroy network for sandbox \"77f789dd2ae82f9cb3145ebd760d0dce00a27a38de683b2799494c90f47eb098\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.946399 containerd[1639]: time="2026-01-21T05:56:07.946195312Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"77f789dd2ae82f9cb3145ebd760d0dce00a27a38de683b2799494c90f47eb098\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.949965 kubelet[2982]: E0121 05:56:07.946621 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"77f789dd2ae82f9cb3145ebd760d0dce00a27a38de683b2799494c90f47eb098\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.949965 kubelet[2982]: E0121 05:56:07.948633 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"77f789dd2ae82f9cb3145ebd760d0dce00a27a38de683b2799494c90f47eb098\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zftdp" Jan 21 05:56:07.950625 kubelet[2982]: E0121 05:56:07.950592 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"77f789dd2ae82f9cb3145ebd760d0dce00a27a38de683b2799494c90f47eb098\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zftdp" Jan 21 05:56:07.953066 kubelet[2982]: E0121 05:56:07.952947 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"77f789dd2ae82f9cb3145ebd760d0dce00a27a38de683b2799494c90f47eb098\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:56:07.956602 containerd[1639]: time="2026-01-21T05:56:07.955178212Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"0da86421731ab4d4495d83e68c9c2c265facde362c89f6502ff713132f1484ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.958010 kubelet[2982]: E0121 05:56:07.955595 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0da86421731ab4d4495d83e68c9c2c265facde362c89f6502ff713132f1484ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.958396 kubelet[2982]: E0121 05:56:07.958360 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0da86421731ab4d4495d83e68c9c2c265facde362c89f6502ff713132f1484ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:56:07.959385 kubelet[2982]: E0121 05:56:07.958510 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0da86421731ab4d4495d83e68c9c2c265facde362c89f6502ff713132f1484ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:56:07.966200 kubelet[2982]: E0121 05:56:07.960632 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-qpdqz_kube-system(9bc2381f-9591-401f-8403-a07aa98276a8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-qpdqz_kube-system(9bc2381f-9591-401f-8403-a07aa98276a8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0da86421731ab4d4495d83e68c9c2c265facde362c89f6502ff713132f1484ad\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-qpdqz" podUID="9bc2381f-9591-401f-8403-a07aa98276a8" Jan 21 05:56:07.967000 containerd[1639]: time="2026-01-21T05:56:07.966590404Z" level=error msg="Failed to destroy network for sandbox \"f2ead182000dfd805a668992e34a7fff4879a8faf04d27676ade62073e8c912f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.969403 containerd[1639]: time="2026-01-21T05:56:07.967191332Z" level=error msg="Failed to destroy network for sandbox \"ec4863cd127040180c06d203b2edb59878bb9f06558c8242bac9786eb80178c9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.987027 containerd[1639]: time="2026-01-21T05:56:07.985077613Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f2ead182000dfd805a668992e34a7fff4879a8faf04d27676ade62073e8c912f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.994242 kubelet[2982]: E0121 05:56:07.994187 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f2ead182000dfd805a668992e34a7fff4879a8faf04d27676ade62073e8c912f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:07.995350 kubelet[2982]: E0121 05:56:07.995251 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f2ead182000dfd805a668992e34a7fff4879a8faf04d27676ade62073e8c912f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:56:07.995350 kubelet[2982]: E0121 05:56:07.995298 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f2ead182000dfd805a668992e34a7fff4879a8faf04d27676ade62073e8c912f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:56:07.996166 kubelet[2982]: E0121 05:56:07.996103 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-sgzh7_kube-system(36de057e-0cdd-42d8-a113-435fdf1d98e9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-sgzh7_kube-system(36de057e-0cdd-42d8-a113-435fdf1d98e9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f2ead182000dfd805a668992e34a7fff4879a8faf04d27676ade62073e8c912f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-sgzh7" podUID="36de057e-0cdd-42d8-a113-435fdf1d98e9" Jan 21 05:56:08.001628 containerd[1639]: time="2026-01-21T05:56:08.001569352Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ec4863cd127040180c06d203b2edb59878bb9f06558c8242bac9786eb80178c9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:08.003008 kubelet[2982]: E0121 05:56:08.002946 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ec4863cd127040180c06d203b2edb59878bb9f06558c8242bac9786eb80178c9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:08.003160 kubelet[2982]: E0121 05:56:08.003131 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ec4863cd127040180c06d203b2edb59878bb9f06558c8242bac9786eb80178c9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:56:08.003384 kubelet[2982]: E0121 05:56:08.003238 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ec4863cd127040180c06d203b2edb59878bb9f06558c8242bac9786eb80178c9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:56:08.003384 kubelet[2982]: E0121 05:56:08.003315 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ec4863cd127040180c06d203b2edb59878bb9f06558c8242bac9786eb80178c9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:56:08.500221 systemd[1]: run-netns-cni\x2dd409b915\x2df27d\x2dc6a8\x2d2801\x2d558cb61622c7.mount: Deactivated successfully. Jan 21 05:56:08.506019 systemd[1]: run-netns-cni\x2d1204f1c2\x2d7b6d\x2d45b2\x2de4a9\x2d852aa8bac066.mount: Deactivated successfully. Jan 21 05:56:08.506132 systemd[1]: run-netns-cni\x2d7718bcd4\x2d112e\x2db009\x2dd21c\x2df1c79d8f1c29.mount: Deactivated successfully. Jan 21 05:56:08.506233 systemd[1]: run-netns-cni\x2d118c8941\x2d26eb\x2dd996\x2ddf03\x2d22acda432763.mount: Deactivated successfully. Jan 21 05:56:10.999000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.61:22-10.0.0.1:50990 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:11.000198 systemd[1]: Started sshd@12-10.0.0.61:22-10.0.0.1:50990.service - OpenSSH per-connection server daemon (10.0.0.1:50990). Jan 21 05:56:11.044018 kernel: audit: type=1130 audit(1768974970.999:602): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.61:22-10.0.0.1:50990 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:11.481526 sshd[4888]: Accepted publickey for core from 10.0.0.1 port 50990 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:11.480000 audit[4888]: USER_ACCT pid=4888 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:11.493387 sshd-session[4888]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:11.514054 kernel: audit: type=1101 audit(1768974971.480:603): pid=4888 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:11.487000 audit[4888]: CRED_ACQ pid=4888 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:11.552612 systemd-logind[1620]: New session 14 of user core. Jan 21 05:56:11.564216 kernel: audit: type=1103 audit(1768974971.487:604): pid=4888 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:11.487000 audit[4888]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff94c75b10 a2=3 a3=0 items=0 ppid=1 pid=4888 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:11.628997 kernel: audit: type=1006 audit(1768974971.487:605): pid=4888 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=14 res=1 Jan 21 05:56:11.629145 kernel: audit: type=1300 audit(1768974971.487:605): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff94c75b10 a2=3 a3=0 items=0 ppid=1 pid=4888 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:11.629188 kernel: audit: type=1327 audit(1768974971.487:605): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:11.487000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:11.641199 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 21 05:56:11.670000 audit[4888]: USER_START pid=4888 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:11.717523 kernel: audit: type=1105 audit(1768974971.670:606): pid=4888 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:11.761324 kernel: audit: type=1103 audit(1768974971.688:607): pid=4892 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:11.688000 audit[4892]: CRED_ACQ pid=4892 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:12.614260 sshd[4892]: Connection closed by 10.0.0.1 port 50990 Jan 21 05:56:12.619287 sshd-session[4888]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:12.649000 audit[4888]: USER_END pid=4888 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:12.662487 systemd[1]: sshd@12-10.0.0.61:22-10.0.0.1:50990.service: Deactivated successfully. Jan 21 05:56:12.690314 kernel: audit: type=1106 audit(1768974972.649:608): pid=4888 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:12.690368 kernel: audit: type=1104 audit(1768974972.649:609): pid=4888 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:12.649000 audit[4888]: CRED_DISP pid=4888 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:12.682300 systemd[1]: session-14.scope: Deactivated successfully. Jan 21 05:56:12.700319 systemd-logind[1620]: Session 14 logged out. Waiting for processes to exit. Jan 21 05:56:12.710473 systemd-logind[1620]: Removed session 14. Jan 21 05:56:12.659000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.61:22-10.0.0.1:50990 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:12.748050 kernel: audit: type=1131 audit(1768974972.659:610): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.61:22-10.0.0.1:50990 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:17.651000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.61:22-10.0.0.1:52162 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:17.652408 systemd[1]: Started sshd@13-10.0.0.61:22-10.0.0.1:52162.service - OpenSSH per-connection server daemon (10.0.0.1:52162). Jan 21 05:56:17.678877 kernel: audit: type=1130 audit(1768974977.651:611): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.61:22-10.0.0.1:52162 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:17.995625 sshd[4908]: Accepted publickey for core from 10.0.0.1 port 52162 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:17.994000 audit[4908]: USER_ACCT pid=4908 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:18.025173 sshd-session[4908]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:18.041262 kernel: audit: type=1101 audit(1768974977.994:612): pid=4908 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:18.012000 audit[4908]: CRED_ACQ pid=4908 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:18.103907 kernel: audit: type=1103 audit(1768974978.012:613): pid=4908 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:18.104308 kernel: audit: type=1006 audit(1768974978.012:614): pid=4908 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=15 res=1 Jan 21 05:56:18.012000 audit[4908]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd55f9b2c0 a2=3 a3=0 items=0 ppid=1 pid=4908 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:18.115582 systemd-logind[1620]: New session 15 of user core. Jan 21 05:56:18.160079 kernel: audit: type=1300 audit(1768974978.012:614): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd55f9b2c0 a2=3 a3=0 items=0 ppid=1 pid=4908 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:18.160508 kernel: audit: type=1327 audit(1768974978.012:614): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:18.012000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:18.170204 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 21 05:56:18.245000 audit[4908]: USER_START pid=4908 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:18.288402 kernel: audit: type=1105 audit(1768974978.245:615): pid=4908 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:18.259000 audit[4913]: CRED_ACQ pid=4913 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:18.365195 kernel: audit: type=1103 audit(1768974978.259:616): pid=4913 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:19.277334 containerd[1639]: time="2026-01-21T05:56:19.276478014Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:56:19.291111 containerd[1639]: time="2026-01-21T05:56:19.287602049Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:56:19.298319 containerd[1639]: time="2026-01-21T05:56:19.298242916Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,}" Jan 21 05:56:19.298772 containerd[1639]: time="2026-01-21T05:56:19.298346791Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,}" Jan 21 05:56:19.585026 sshd[4913]: Connection closed by 10.0.0.1 port 52162 Jan 21 05:56:19.584602 sshd-session[4908]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:19.602000 audit[4908]: USER_END pid=4908 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:19.661843 kernel: audit: type=1106 audit(1768974979.602:617): pid=4908 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:19.666763 systemd-logind[1620]: Session 15 logged out. Waiting for processes to exit. Jan 21 05:56:19.668111 systemd[1]: sshd@13-10.0.0.61:22-10.0.0.1:52162.service: Deactivated successfully. Jan 21 05:56:19.654000 audit[4908]: CRED_DISP pid=4908 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:19.692196 systemd[1]: session-15.scope: Deactivated successfully. Jan 21 05:56:19.705528 systemd-logind[1620]: Removed session 15. Jan 21 05:56:19.719835 kernel: audit: type=1104 audit(1768974979.654:618): pid=4908 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:19.666000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.61:22-10.0.0.1:52162 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:20.083939 containerd[1639]: time="2026-01-21T05:56:20.083581509Z" level=error msg="Failed to destroy network for sandbox \"47333cbe1f483d7f16e8de78488ac4e4baae2a43afe406f2b09d347a05453d0e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.095180 containerd[1639]: time="2026-01-21T05:56:20.094910089Z" level=error msg="Failed to destroy network for sandbox \"9a406226cb74f17c5ee8cd36eef279447795e7d39cd489cb3c4e91f081e4e3f0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.108143 containerd[1639]: time="2026-01-21T05:56:20.108005854Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"47333cbe1f483d7f16e8de78488ac4e4baae2a43afe406f2b09d347a05453d0e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.139139 systemd[1]: run-netns-cni\x2dc23b86c7\x2d3e0c\x2d0b34\x2de618\x2d2e728bb01f75.mount: Deactivated successfully. Jan 21 05:56:20.153772 kubelet[2982]: E0121 05:56:20.153155 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47333cbe1f483d7f16e8de78488ac4e4baae2a43afe406f2b09d347a05453d0e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.153772 kubelet[2982]: E0121 05:56:20.153584 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47333cbe1f483d7f16e8de78488ac4e4baae2a43afe406f2b09d347a05453d0e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:56:20.153772 kubelet[2982]: E0121 05:56:20.153617 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47333cbe1f483d7f16e8de78488ac4e4baae2a43afe406f2b09d347a05453d0e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-m7pjj" Jan 21 05:56:20.157921 kubelet[2982]: E0121 05:56:20.157524 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"47333cbe1f483d7f16e8de78488ac4e4baae2a43afe406f2b09d347a05453d0e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:56:20.167239 containerd[1639]: time="2026-01-21T05:56:20.167126469Z" level=error msg="Failed to destroy network for sandbox \"575f520b1000ef86309f9d0127cac09c9e5cc0d1a633ba7fc6b7b2a9d70d14fb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.170474 containerd[1639]: time="2026-01-21T05:56:20.170427331Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9a406226cb74f17c5ee8cd36eef279447795e7d39cd489cb3c4e91f081e4e3f0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.182583 containerd[1639]: time="2026-01-21T05:56:20.182466263Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"575f520b1000ef86309f9d0127cac09c9e5cc0d1a633ba7fc6b7b2a9d70d14fb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.186465 kubelet[2982]: E0121 05:56:20.186348 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"575f520b1000ef86309f9d0127cac09c9e5cc0d1a633ba7fc6b7b2a9d70d14fb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.186465 kubelet[2982]: E0121 05:56:20.186429 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"575f520b1000ef86309f9d0127cac09c9e5cc0d1a633ba7fc6b7b2a9d70d14fb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:56:20.186465 kubelet[2982]: E0121 05:56:20.186457 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"575f520b1000ef86309f9d0127cac09c9e5cc0d1a633ba7fc6b7b2a9d70d14fb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" Jan 21 05:56:20.186821 kubelet[2982]: E0121 05:56:20.186518 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"575f520b1000ef86309f9d0127cac09c9e5cc0d1a633ba7fc6b7b2a9d70d14fb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:56:20.186821 kubelet[2982]: E0121 05:56:20.186584 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9a406226cb74f17c5ee8cd36eef279447795e7d39cd489cb3c4e91f081e4e3f0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.186821 kubelet[2982]: E0121 05:56:20.186612 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9a406226cb74f17c5ee8cd36eef279447795e7d39cd489cb3c4e91f081e4e3f0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:56:20.187206 kubelet[2982]: E0121 05:56:20.186820 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9a406226cb74f17c5ee8cd36eef279447795e7d39cd489cb3c4e91f081e4e3f0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" Jan 21 05:56:20.187314 kubelet[2982]: E0121 05:56:20.187267 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9a406226cb74f17c5ee8cd36eef279447795e7d39cd489cb3c4e91f081e4e3f0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:56:20.214025 kubelet[2982]: E0121 05:56:20.213978 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:20.222951 containerd[1639]: time="2026-01-21T05:56:20.222910145Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,}" Jan 21 05:56:20.226810 containerd[1639]: time="2026-01-21T05:56:20.225194850Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-db7d6575b-rxmxq,Uid:5efe9c76-b287-463e-ba8c-614903640526,Namespace:calico-system,Attempt:0,}" Jan 21 05:56:20.255778 containerd[1639]: time="2026-01-21T05:56:20.251944371Z" level=error msg="Failed to destroy network for sandbox \"187c085c067b2def3f8eece868e8ea55364a25063a454a74e71d45e93ce63ff0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.281548 containerd[1639]: time="2026-01-21T05:56:20.280624646Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"187c085c067b2def3f8eece868e8ea55364a25063a454a74e71d45e93ce63ff0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.288805 kubelet[2982]: E0121 05:56:20.287910 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"187c085c067b2def3f8eece868e8ea55364a25063a454a74e71d45e93ce63ff0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.292590 kubelet[2982]: E0121 05:56:20.292474 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"187c085c067b2def3f8eece868e8ea55364a25063a454a74e71d45e93ce63ff0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zftdp" Jan 21 05:56:20.292590 kubelet[2982]: E0121 05:56:20.292524 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"187c085c067b2def3f8eece868e8ea55364a25063a454a74e71d45e93ce63ff0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zftdp" Jan 21 05:56:20.293582 kubelet[2982]: E0121 05:56:20.293273 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"187c085c067b2def3f8eece868e8ea55364a25063a454a74e71d45e93ce63ff0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:56:20.489575 systemd[1]: run-netns-cni\x2dead1ac1a\x2d287a\x2d91e7\x2d18c1\x2d1ebe3f3b4936.mount: Deactivated successfully. Jan 21 05:56:20.490136 systemd[1]: run-netns-cni\x2de2d065d9\x2d6391\x2d2bfc\x2d43e1\x2d822cc16581b9.mount: Deactivated successfully. Jan 21 05:56:20.490266 systemd[1]: run-netns-cni\x2db6496cac\x2dd7d4\x2d4b6b\x2d79c3\x2d17fb72f036cd.mount: Deactivated successfully. Jan 21 05:56:20.571709 containerd[1639]: time="2026-01-21T05:56:20.570542819Z" level=error msg="Failed to destroy network for sandbox \"c10aea6333802e48abb48f2fbc68b966970a374ce5072abddf4af016d040adc0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.583312 systemd[1]: run-netns-cni\x2db70db052\x2d9f4c\x2d7165\x2d839e\x2dbec166c5b00f.mount: Deactivated successfully. Jan 21 05:56:20.593590 containerd[1639]: time="2026-01-21T05:56:20.592030973Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-db7d6575b-rxmxq,Uid:5efe9c76-b287-463e-ba8c-614903640526,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c10aea6333802e48abb48f2fbc68b966970a374ce5072abddf4af016d040adc0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.595888 kubelet[2982]: E0121 05:56:20.594448 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c10aea6333802e48abb48f2fbc68b966970a374ce5072abddf4af016d040adc0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.595888 kubelet[2982]: E0121 05:56:20.594546 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c10aea6333802e48abb48f2fbc68b966970a374ce5072abddf4af016d040adc0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:56:20.595888 kubelet[2982]: E0121 05:56:20.594585 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c10aea6333802e48abb48f2fbc68b966970a374ce5072abddf4af016d040adc0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-db7d6575b-rxmxq" Jan 21 05:56:20.596193 kubelet[2982]: E0121 05:56:20.594802 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-db7d6575b-rxmxq_calico-system(5efe9c76-b287-463e-ba8c-614903640526)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-db7d6575b-rxmxq_calico-system(5efe9c76-b287-463e-ba8c-614903640526)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c10aea6333802e48abb48f2fbc68b966970a374ce5072abddf4af016d040adc0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-db7d6575b-rxmxq" podUID="5efe9c76-b287-463e-ba8c-614903640526" Jan 21 05:56:20.655043 containerd[1639]: time="2026-01-21T05:56:20.652295088Z" level=error msg="Failed to destroy network for sandbox \"7a61314f26c716d29f8a3cb114deee7e8da1a0cfd891e58fc3e461ec08f22592\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.662174 systemd[1]: run-netns-cni\x2d82910581\x2d08fb\x2d7aa1\x2d943f\x2d2ad5f9ad9576.mount: Deactivated successfully. Jan 21 05:56:20.677262 containerd[1639]: time="2026-01-21T05:56:20.676751912Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7a61314f26c716d29f8a3cb114deee7e8da1a0cfd891e58fc3e461ec08f22592\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.677525 kubelet[2982]: E0121 05:56:20.677264 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7a61314f26c716d29f8a3cb114deee7e8da1a0cfd891e58fc3e461ec08f22592\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:20.677525 kubelet[2982]: E0121 05:56:20.677354 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7a61314f26c716d29f8a3cb114deee7e8da1a0cfd891e58fc3e461ec08f22592\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:56:20.677525 kubelet[2982]: E0121 05:56:20.677386 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7a61314f26c716d29f8a3cb114deee7e8da1a0cfd891e58fc3e461ec08f22592\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-sgzh7" Jan 21 05:56:20.677898 kubelet[2982]: E0121 05:56:20.677451 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-sgzh7_kube-system(36de057e-0cdd-42d8-a113-435fdf1d98e9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-sgzh7_kube-system(36de057e-0cdd-42d8-a113-435fdf1d98e9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7a61314f26c716d29f8a3cb114deee7e8da1a0cfd891e58fc3e461ec08f22592\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-sgzh7" podUID="36de057e-0cdd-42d8-a113-435fdf1d98e9" Jan 21 05:56:21.213000 kubelet[2982]: E0121 05:56:21.210044 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:21.230804 containerd[1639]: time="2026-01-21T05:56:21.230468900Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,}" Jan 21 05:56:21.611952 containerd[1639]: time="2026-01-21T05:56:21.607534962Z" level=error msg="Failed to destroy network for sandbox \"04cf12d4325fd979070ec818caf8d72a2ae30bf8c5a5a07f5a85800e4fd2f25c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:21.612626 systemd[1]: run-netns-cni\x2d03e793a1\x2db8d3\x2d7143\x2d5428\x2d9098542f8f01.mount: Deactivated successfully. Jan 21 05:56:21.628757 containerd[1639]: time="2026-01-21T05:56:21.628356385Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"04cf12d4325fd979070ec818caf8d72a2ae30bf8c5a5a07f5a85800e4fd2f25c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:21.638957 kubelet[2982]: E0121 05:56:21.636987 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"04cf12d4325fd979070ec818caf8d72a2ae30bf8c5a5a07f5a85800e4fd2f25c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:21.638957 kubelet[2982]: E0121 05:56:21.637245 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"04cf12d4325fd979070ec818caf8d72a2ae30bf8c5a5a07f5a85800e4fd2f25c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:56:21.638957 kubelet[2982]: E0121 05:56:21.637280 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"04cf12d4325fd979070ec818caf8d72a2ae30bf8c5a5a07f5a85800e4fd2f25c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-qpdqz" Jan 21 05:56:21.639404 kubelet[2982]: E0121 05:56:21.637350 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-qpdqz_kube-system(9bc2381f-9591-401f-8403-a07aa98276a8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-qpdqz_kube-system(9bc2381f-9591-401f-8403-a07aa98276a8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"04cf12d4325fd979070ec818caf8d72a2ae30bf8c5a5a07f5a85800e4fd2f25c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-qpdqz" podUID="9bc2381f-9591-401f-8403-a07aa98276a8" Jan 21 05:56:22.213604 containerd[1639]: time="2026-01-21T05:56:22.213283128Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:56:22.512408 containerd[1639]: time="2026-01-21T05:56:22.511077402Z" level=error msg="Failed to destroy network for sandbox \"258bf2490448595bc61465304521d3dbf12e9c1349d754e6df9216968aaf79c3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:22.519598 systemd[1]: run-netns-cni\x2d14ad3ef7\x2d323f\x2d820c\x2df956\x2d61148201f8a9.mount: Deactivated successfully. Jan 21 05:56:22.555817 containerd[1639]: time="2026-01-21T05:56:22.555583189Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"258bf2490448595bc61465304521d3dbf12e9c1349d754e6df9216968aaf79c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:22.560832 kubelet[2982]: E0121 05:56:22.559004 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"258bf2490448595bc61465304521d3dbf12e9c1349d754e6df9216968aaf79c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:22.560832 kubelet[2982]: E0121 05:56:22.559218 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"258bf2490448595bc61465304521d3dbf12e9c1349d754e6df9216968aaf79c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:56:22.560832 kubelet[2982]: E0121 05:56:22.559252 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"258bf2490448595bc61465304521d3dbf12e9c1349d754e6df9216968aaf79c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" Jan 21 05:56:22.561540 kubelet[2982]: E0121 05:56:22.559325 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"258bf2490448595bc61465304521d3dbf12e9c1349d754e6df9216968aaf79c3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:56:23.324551 containerd[1639]: time="2026-01-21T05:56:23.321613357Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,}" Jan 21 05:56:23.780625 containerd[1639]: time="2026-01-21T05:56:23.780409376Z" level=error msg="Failed to destroy network for sandbox \"f99dedd4a964066f1fd453973b965a995d96b14cf370d4915de2c3b3ba06b336\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:23.789364 systemd[1]: run-netns-cni\x2d0f2b82a2\x2df393\x2dd967\x2d74b6\x2d16c57b371282.mount: Deactivated successfully. Jan 21 05:56:23.832296 containerd[1639]: time="2026-01-21T05:56:23.830513973Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f99dedd4a964066f1fd453973b965a995d96b14cf370d4915de2c3b3ba06b336\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:23.834770 kubelet[2982]: E0121 05:56:23.834483 2982 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f99dedd4a964066f1fd453973b965a995d96b14cf370d4915de2c3b3ba06b336\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 21 05:56:23.834770 kubelet[2982]: E0121 05:56:23.834630 2982 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f99dedd4a964066f1fd453973b965a995d96b14cf370d4915de2c3b3ba06b336\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:56:23.836527 kubelet[2982]: E0121 05:56:23.834794 2982 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f99dedd4a964066f1fd453973b965a995d96b14cf370d4915de2c3b3ba06b336\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" Jan 21 05:56:23.836527 kubelet[2982]: E0121 05:56:23.834859 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f99dedd4a964066f1fd453973b965a995d96b14cf370d4915de2c3b3ba06b336\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:56:24.230511 kubelet[2982]: E0121 05:56:24.225981 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:24.623000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.61:22-10.0.0.1:42614 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:24.624326 systemd[1]: Started sshd@14-10.0.0.61:22-10.0.0.1:42614.service - OpenSSH per-connection server daemon (10.0.0.1:42614). Jan 21 05:56:24.673230 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:56:24.673979 kernel: audit: type=1130 audit(1768974984.623:620): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.61:22-10.0.0.1:42614 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:24.975000 audit[5215]: USER_ACCT pid=5215 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:24.979983 sshd[5215]: Accepted publickey for core from 10.0.0.1 port 42614 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:25.002559 sshd-session[5215]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:25.005105 kernel: audit: type=1101 audit(1768974984.975:621): pid=5215 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:24.985000 audit[5215]: CRED_ACQ pid=5215 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.037034 kernel: audit: type=1103 audit(1768974984.985:622): pid=5215 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.060207 kernel: audit: type=1006 audit(1768974984.986:623): pid=5215 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=16 res=1 Jan 21 05:56:25.061492 systemd-logind[1620]: New session 16 of user core. Jan 21 05:56:24.986000 audit[5215]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdfa618670 a2=3 a3=0 items=0 ppid=1 pid=5215 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:25.122920 kernel: audit: type=1300 audit(1768974984.986:623): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdfa618670 a2=3 a3=0 items=0 ppid=1 pid=5215 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:24.986000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:25.125855 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 21 05:56:25.138945 kernel: audit: type=1327 audit(1768974984.986:623): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:25.140000 audit[5215]: USER_START pid=5215 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.149000 audit[5219]: CRED_ACQ pid=5219 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.198964 kernel: audit: type=1105 audit(1768974985.140:624): pid=5215 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.199093 kernel: audit: type=1103 audit(1768974985.149:625): pid=5219 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.639356 sshd[5219]: Connection closed by 10.0.0.1 port 42614 Jan 21 05:56:25.640318 sshd-session[5215]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:25.643000 audit[5215]: USER_END pid=5215 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.652855 systemd[1]: sshd@14-10.0.0.61:22-10.0.0.1:42614.service: Deactivated successfully. Jan 21 05:56:25.661820 systemd[1]: session-16.scope: Deactivated successfully. Jan 21 05:56:25.670013 systemd-logind[1620]: Session 16 logged out. Waiting for processes to exit. Jan 21 05:56:25.672448 systemd-logind[1620]: Removed session 16. Jan 21 05:56:25.643000 audit[5215]: CRED_DISP pid=5215 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.729046 kernel: audit: type=1106 audit(1768974985.643:626): pid=5215 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.731298 kernel: audit: type=1104 audit(1768974985.643:627): pid=5215 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:25.655000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.61:22-10.0.0.1:42614 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:26.578478 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount93539148.mount: Deactivated successfully. Jan 21 05:56:26.781586 containerd[1639]: time="2026-01-21T05:56:26.778070191Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:56:26.797378 containerd[1639]: time="2026-01-21T05:56:26.796140187Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156880025" Jan 21 05:56:26.838761 containerd[1639]: time="2026-01-21T05:56:26.835318630Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:56:26.886039 containerd[1639]: time="2026-01-21T05:56:26.885897623Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 21 05:56:26.889789 containerd[1639]: time="2026-01-21T05:56:26.887944207Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 1m1.392064027s" Jan 21 05:56:26.889789 containerd[1639]: time="2026-01-21T05:56:26.888003048Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Jan 21 05:56:27.096413 containerd[1639]: time="2026-01-21T05:56:27.094773257Z" level=info msg="CreateContainer within sandbox \"e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jan 21 05:56:27.178982 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3224007032.mount: Deactivated successfully. Jan 21 05:56:27.180451 containerd[1639]: time="2026-01-21T05:56:27.180314179Z" level=info msg="Container 98cadf0e1f21dbbbf2727d9fac57c15f589062e9dc0bea52d06636c8808e92f0: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:56:27.228240 containerd[1639]: time="2026-01-21T05:56:27.227979325Z" level=info msg="CreateContainer within sandbox \"e06762d872d118abf8f3982353b0f45a216ecf201c12c5b12d4c85a2148d845c\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"98cadf0e1f21dbbbf2727d9fac57c15f589062e9dc0bea52d06636c8808e92f0\"" Jan 21 05:56:27.235155 containerd[1639]: time="2026-01-21T05:56:27.231573721Z" level=info msg="StartContainer for \"98cadf0e1f21dbbbf2727d9fac57c15f589062e9dc0bea52d06636c8808e92f0\"" Jan 21 05:56:27.255503 containerd[1639]: time="2026-01-21T05:56:27.254382369Z" level=info msg="connecting to shim 98cadf0e1f21dbbbf2727d9fac57c15f589062e9dc0bea52d06636c8808e92f0" address="unix:///run/containerd/s/a097f43dd299e2930c4a70b814880e743e86b92e44c5686bf5acf59179f1242e" protocol=ttrpc version=3 Jan 21 05:56:27.425827 systemd[1]: Started cri-containerd-98cadf0e1f21dbbbf2727d9fac57c15f589062e9dc0bea52d06636c8808e92f0.scope - libcontainer container 98cadf0e1f21dbbbf2727d9fac57c15f589062e9dc0bea52d06636c8808e92f0. Jan 21 05:56:27.667000 audit: BPF prog-id=170 op=LOAD Jan 21 05:56:27.667000 audit[5237]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0000fc488 a2=98 a3=0 items=0 ppid=3542 pid=5237 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:27.667000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938636164663065316632316462626266323732376439666163353763 Jan 21 05:56:27.669000 audit: BPF prog-id=171 op=LOAD Jan 21 05:56:27.669000 audit[5237]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0000fc218 a2=98 a3=0 items=0 ppid=3542 pid=5237 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:27.669000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938636164663065316632316462626266323732376439666163353763 Jan 21 05:56:27.669000 audit: BPF prog-id=171 op=UNLOAD Jan 21 05:56:27.669000 audit[5237]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=5237 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:27.669000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938636164663065316632316462626266323732376439666163353763 Jan 21 05:56:27.669000 audit: BPF prog-id=170 op=UNLOAD Jan 21 05:56:27.669000 audit[5237]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3542 pid=5237 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:27.669000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938636164663065316632316462626266323732376439666163353763 Jan 21 05:56:27.669000 audit: BPF prog-id=172 op=LOAD Jan 21 05:56:27.669000 audit[5237]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0000fc6e8 a2=98 a3=0 items=0 ppid=3542 pid=5237 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:27.669000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938636164663065316632316462626266323732376439666163353763 Jan 21 05:56:27.798567 containerd[1639]: time="2026-01-21T05:56:27.798420967Z" level=info msg="StartContainer for \"98cadf0e1f21dbbbf2727d9fac57c15f589062e9dc0bea52d06636c8808e92f0\" returns successfully" Jan 21 05:56:28.377148 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jan 21 05:56:28.377439 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jan 21 05:56:28.534595 kubelet[2982]: E0121 05:56:28.534495 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:28.620126 kubelet[2982]: I0121 05:56:28.619818 2982 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-tscdz" podStartSLOduration=4.687352205 podStartE2EDuration="1m55.619792772s" podCreationTimestamp="2026-01-21 05:54:33 +0000 UTC" firstStartedPulling="2026-01-21 05:54:35.992086989 +0000 UTC m=+70.201736692" lastFinishedPulling="2026-01-21 05:56:26.924527556 +0000 UTC m=+181.134177259" observedRunningTime="2026-01-21 05:56:28.610762931 +0000 UTC m=+182.820412935" watchObservedRunningTime="2026-01-21 05:56:28.619792772 +0000 UTC m=+182.829442505" Jan 21 05:56:29.086947 kubelet[2982]: I0121 05:56:29.086765 2982 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5efe9c76-b287-463e-ba8c-614903640526-whisker-backend-key-pair\") pod \"5efe9c76-b287-463e-ba8c-614903640526\" (UID: \"5efe9c76-b287-463e-ba8c-614903640526\") " Jan 21 05:56:29.086947 kubelet[2982]: I0121 05:56:29.086890 2982 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5efe9c76-b287-463e-ba8c-614903640526-whisker-ca-bundle\") pod \"5efe9c76-b287-463e-ba8c-614903640526\" (UID: \"5efe9c76-b287-463e-ba8c-614903640526\") " Jan 21 05:56:29.086947 kubelet[2982]: I0121 05:56:29.086910 2982 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rrl5\" (UniqueName: \"kubernetes.io/projected/5efe9c76-b287-463e-ba8c-614903640526-kube-api-access-4rrl5\") pod \"5efe9c76-b287-463e-ba8c-614903640526\" (UID: \"5efe9c76-b287-463e-ba8c-614903640526\") " Jan 21 05:56:29.097377 kubelet[2982]: I0121 05:56:29.097265 2982 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5efe9c76-b287-463e-ba8c-614903640526-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "5efe9c76-b287-463e-ba8c-614903640526" (UID: "5efe9c76-b287-463e-ba8c-614903640526"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jan 21 05:56:29.131091 kubelet[2982]: I0121 05:56:29.130982 2982 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5efe9c76-b287-463e-ba8c-614903640526-kube-api-access-4rrl5" (OuterVolumeSpecName: "kube-api-access-4rrl5") pod "5efe9c76-b287-463e-ba8c-614903640526" (UID: "5efe9c76-b287-463e-ba8c-614903640526"). InnerVolumeSpecName "kube-api-access-4rrl5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jan 21 05:56:29.134349 systemd[1]: var-lib-kubelet-pods-5efe9c76\x2db287\x2d463e\x2dba8c\x2d614903640526-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d4rrl5.mount: Deactivated successfully. Jan 21 05:56:29.134531 systemd[1]: var-lib-kubelet-pods-5efe9c76\x2db287\x2d463e\x2dba8c\x2d614903640526-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Jan 21 05:56:29.137507 kubelet[2982]: I0121 05:56:29.136803 2982 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5efe9c76-b287-463e-ba8c-614903640526-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "5efe9c76-b287-463e-ba8c-614903640526" (UID: "5efe9c76-b287-463e-ba8c-614903640526"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Jan 21 05:56:29.189161 kubelet[2982]: I0121 05:56:29.189055 2982 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5efe9c76-b287-463e-ba8c-614903640526-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Jan 21 05:56:29.189161 kubelet[2982]: I0121 05:56:29.189112 2982 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5efe9c76-b287-463e-ba8c-614903640526-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Jan 21 05:56:29.189161 kubelet[2982]: I0121 05:56:29.189127 2982 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4rrl5\" (UniqueName: \"kubernetes.io/projected/5efe9c76-b287-463e-ba8c-614903640526-kube-api-access-4rrl5\") on node \"localhost\" DevicePath \"\"" Jan 21 05:56:29.548182 kubelet[2982]: E0121 05:56:29.548060 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:29.596099 systemd[1]: Removed slice kubepods-besteffort-pod5efe9c76_b287_463e_ba8c_614903640526.slice - libcontainer container kubepods-besteffort-pod5efe9c76_b287_463e_ba8c_614903640526.slice. Jan 21 05:56:30.026110 systemd[1]: Created slice kubepods-besteffort-podda0d7215_59d8_4e4a_9787_c2942086bde3.slice - libcontainer container kubepods-besteffort-podda0d7215_59d8_4e4a_9787_c2942086bde3.slice. Jan 21 05:56:30.108949 kubelet[2982]: I0121 05:56:30.108414 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/da0d7215-59d8-4e4a-9787-c2942086bde3-whisker-backend-key-pair\") pod \"whisker-cd95c566-k55c2\" (UID: \"da0d7215-59d8-4e4a-9787-c2942086bde3\") " pod="calico-system/whisker-cd95c566-k55c2" Jan 21 05:56:30.108949 kubelet[2982]: I0121 05:56:30.108484 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da0d7215-59d8-4e4a-9787-c2942086bde3-whisker-ca-bundle\") pod \"whisker-cd95c566-k55c2\" (UID: \"da0d7215-59d8-4e4a-9787-c2942086bde3\") " pod="calico-system/whisker-cd95c566-k55c2" Jan 21 05:56:30.108949 kubelet[2982]: I0121 05:56:30.108518 2982 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgkt9\" (UniqueName: \"kubernetes.io/projected/da0d7215-59d8-4e4a-9787-c2942086bde3-kube-api-access-rgkt9\") pod \"whisker-cd95c566-k55c2\" (UID: \"da0d7215-59d8-4e4a-9787-c2942086bde3\") " pod="calico-system/whisker-cd95c566-k55c2" Jan 21 05:56:30.243838 kubelet[2982]: I0121 05:56:30.243608 2982 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5efe9c76-b287-463e-ba8c-614903640526" path="/var/lib/kubelet/pods/5efe9c76-b287-463e-ba8c-614903640526/volumes" Jan 21 05:56:30.343805 containerd[1639]: time="2026-01-21T05:56:30.342125739Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-cd95c566-k55c2,Uid:da0d7215-59d8-4e4a-9787-c2942086bde3,Namespace:calico-system,Attempt:0,}" Jan 21 05:56:30.702000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.61:22-10.0.0.1:42618 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:30.702986 systemd[1]: Started sshd@15-10.0.0.61:22-10.0.0.1:42618.service - OpenSSH per-connection server daemon (10.0.0.1:42618). Jan 21 05:56:30.715192 kernel: kauditd_printk_skb: 16 callbacks suppressed Jan 21 05:56:30.716055 kernel: audit: type=1130 audit(1768974990.702:634): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.61:22-10.0.0.1:42618 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:30.968000 audit[5370]: USER_ACCT pid=5370 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:30.974824 sshd[5370]: Accepted publickey for core from 10.0.0.1 port 42618 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:30.973942 sshd-session[5370]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:30.997465 systemd-logind[1620]: New session 17 of user core. Jan 21 05:56:31.017986 kernel: audit: type=1101 audit(1768974990.968:635): pid=5370 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:30.970000 audit[5370]: CRED_ACQ pid=5370 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:31.063930 kernel: audit: type=1103 audit(1768974990.970:636): pid=5370 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:31.064080 kernel: audit: type=1006 audit(1768974990.970:637): pid=5370 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=17 res=1 Jan 21 05:56:31.064127 kernel: audit: type=1300 audit(1768974990.970:637): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd1fe3b6b0 a2=3 a3=0 items=0 ppid=1 pid=5370 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:30.970000 audit[5370]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd1fe3b6b0 a2=3 a3=0 items=0 ppid=1 pid=5370 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:31.066018 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 21 05:56:31.105128 kernel: audit: type=1327 audit(1768974990.970:637): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:30.970000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:31.085000 audit[5370]: USER_START pid=5370 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:31.152951 kernel: audit: type=1105 audit(1768974991.085:638): pid=5370 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:31.095000 audit[5384]: CRED_ACQ pid=5384 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:31.188150 kernel: audit: type=1103 audit(1768974991.095:639): pid=5384 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:31.214966 containerd[1639]: time="2026-01-21T05:56:31.214878492Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:56:31.629738 sshd[5384]: Connection closed by 10.0.0.1 port 42618 Jan 21 05:56:31.630981 sshd-session[5370]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:31.660000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.61:22-10.0.0.1:42622 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:31.660525 systemd[1]: Started sshd@16-10.0.0.61:22-10.0.0.1:42622.service - OpenSSH per-connection server daemon (10.0.0.1:42622). Jan 21 05:56:31.681505 systemd[1]: sshd@15-10.0.0.61:22-10.0.0.1:42618.service: Deactivated successfully. Jan 21 05:56:31.691421 kernel: audit: type=1130 audit(1768974991.660:640): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.61:22-10.0.0.1:42622 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:31.686969 systemd-networkd[1528]: cali3b001fe5b05: Link UP Jan 21 05:56:31.688890 systemd[1]: session-17.scope: Deactivated successfully. Jan 21 05:56:31.754957 kernel: audit: type=1106 audit(1768974991.673:641): pid=5370 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:31.673000 audit[5370]: USER_END pid=5370 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:31.707939 systemd-networkd[1528]: cali3b001fe5b05: Gained carrier Jan 21 05:56:31.712188 systemd-logind[1620]: Session 17 logged out. Waiting for processes to exit. Jan 21 05:56:31.748880 systemd-logind[1620]: Removed session 17. Jan 21 05:56:31.673000 audit[5370]: CRED_DISP pid=5370 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:31.681000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.61:22-10.0.0.1:42618 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:31.808078 containerd[1639]: 2026-01-21 05:56:30.498 [INFO][5350] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 21 05:56:31.808078 containerd[1639]: 2026-01-21 05:56:30.664 [INFO][5350] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--cd95c566--k55c2-eth0 whisker-cd95c566- calico-system da0d7215-59d8-4e4a-9787-c2942086bde3 1392 0 2026-01-21 05:56:29 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:cd95c566 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-cd95c566-k55c2 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali3b001fe5b05 [] [] }} ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Namespace="calico-system" Pod="whisker-cd95c566-k55c2" WorkloadEndpoint="localhost-k8s-whisker--cd95c566--k55c2-" Jan 21 05:56:31.808078 containerd[1639]: 2026-01-21 05:56:30.664 [INFO][5350] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Namespace="calico-system" Pod="whisker-cd95c566-k55c2" WorkloadEndpoint="localhost-k8s-whisker--cd95c566--k55c2-eth0" Jan 21 05:56:31.808078 containerd[1639]: 2026-01-21 05:56:31.168 [INFO][5372] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" HandleID="k8s-pod-network.4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Workload="localhost-k8s-whisker--cd95c566--k55c2-eth0" Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.170 [INFO][5372] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" HandleID="k8s-pod-network.4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Workload="localhost-k8s-whisker--cd95c566--k55c2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002fc0d0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-cd95c566-k55c2", "timestamp":"2026-01-21 05:56:31.168033258 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.170 [INFO][5372] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.170 [INFO][5372] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.171 [INFO][5372] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.249 [INFO][5372] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" host="localhost" Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.307 [INFO][5372] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.358 [INFO][5372] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.380 [INFO][5372] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.392 [INFO][5372] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:31.810369 containerd[1639]: 2026-01-21 05:56:31.393 [INFO][5372] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" host="localhost" Jan 21 05:56:31.811519 containerd[1639]: 2026-01-21 05:56:31.406 [INFO][5372] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a Jan 21 05:56:31.811519 containerd[1639]: 2026-01-21 05:56:31.441 [INFO][5372] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" host="localhost" Jan 21 05:56:31.811519 containerd[1639]: 2026-01-21 05:56:31.478 [INFO][5372] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" host="localhost" Jan 21 05:56:31.811519 containerd[1639]: 2026-01-21 05:56:31.482 [INFO][5372] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" host="localhost" Jan 21 05:56:31.811519 containerd[1639]: 2026-01-21 05:56:31.482 [INFO][5372] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 21 05:56:31.811519 containerd[1639]: 2026-01-21 05:56:31.483 [INFO][5372] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" HandleID="k8s-pod-network.4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Workload="localhost-k8s-whisker--cd95c566--k55c2-eth0" Jan 21 05:56:31.812145 containerd[1639]: 2026-01-21 05:56:31.525 [INFO][5350] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Namespace="calico-system" Pod="whisker-cd95c566-k55c2" WorkloadEndpoint="localhost-k8s-whisker--cd95c566--k55c2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--cd95c566--k55c2-eth0", GenerateName:"whisker-cd95c566-", Namespace:"calico-system", SelfLink:"", UID:"da0d7215-59d8-4e4a-9787-c2942086bde3", ResourceVersion:"1392", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 56, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"cd95c566", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-cd95c566-k55c2", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali3b001fe5b05", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:31.812145 containerd[1639]: 2026-01-21 05:56:31.525 [INFO][5350] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Namespace="calico-system" Pod="whisker-cd95c566-k55c2" WorkloadEndpoint="localhost-k8s-whisker--cd95c566--k55c2-eth0" Jan 21 05:56:31.813182 containerd[1639]: 2026-01-21 05:56:31.525 [INFO][5350] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3b001fe5b05 ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Namespace="calico-system" Pod="whisker-cd95c566-k55c2" WorkloadEndpoint="localhost-k8s-whisker--cd95c566--k55c2-eth0" Jan 21 05:56:31.813182 containerd[1639]: 2026-01-21 05:56:31.698 [INFO][5350] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Namespace="calico-system" Pod="whisker-cd95c566-k55c2" WorkloadEndpoint="localhost-k8s-whisker--cd95c566--k55c2-eth0" Jan 21 05:56:31.813805 containerd[1639]: 2026-01-21 05:56:31.725 [INFO][5350] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Namespace="calico-system" Pod="whisker-cd95c566-k55c2" WorkloadEndpoint="localhost-k8s-whisker--cd95c566--k55c2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--cd95c566--k55c2-eth0", GenerateName:"whisker-cd95c566-", Namespace:"calico-system", SelfLink:"", UID:"da0d7215-59d8-4e4a-9787-c2942086bde3", ResourceVersion:"1392", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 56, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"cd95c566", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a", Pod:"whisker-cd95c566-k55c2", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali3b001fe5b05", MAC:"ae:75:f7:0b:4e:b4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:31.814362 containerd[1639]: 2026-01-21 05:56:31.787 [INFO][5350] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" Namespace="calico-system" Pod="whisker-cd95c566-k55c2" WorkloadEndpoint="localhost-k8s-whisker--cd95c566--k55c2-eth0" Jan 21 05:56:32.068989 sshd[5447]: Accepted publickey for core from 10.0.0.1 port 42622 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:32.067000 audit[5447]: USER_ACCT pid=5447 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:32.087000 audit[5447]: CRED_ACQ pid=5447 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:32.090000 audit[5447]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe511da070 a2=3 a3=0 items=0 ppid=1 pid=5447 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:32.090000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:32.111856 sshd-session[5447]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:32.152588 systemd-logind[1620]: New session 18 of user core. Jan 21 05:56:32.162073 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 21 05:56:32.194000 audit[5447]: USER_START pid=5447 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:32.204000 audit[5551]: CRED_ACQ pid=5551 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:32.225226 containerd[1639]: time="2026-01-21T05:56:32.225136941Z" level=info msg="connecting to shim 4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a" address="unix:///run/containerd/s/d8538daded00b68f6b518ea997a0d7f916ccb29fad9d9596d4dc05ba0c148aec" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:56:32.470946 systemd-networkd[1528]: cali6cfc639d9e3: Link UP Jan 21 05:56:32.474143 systemd-networkd[1528]: cali6cfc639d9e3: Gained carrier Jan 21 05:56:32.492914 systemd[1]: Started cri-containerd-4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a.scope - libcontainer container 4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a. Jan 21 05:56:32.581620 containerd[1639]: 2026-01-21 05:56:31.396 [INFO][5396] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 21 05:56:32.581620 containerd[1639]: 2026-01-21 05:56:31.474 [INFO][5396] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--796779bb--f2fds-eth0 calico-apiserver-796779bb- calico-apiserver dac6efad-1039-4242-b1c8-782ac61e5470 1091 0 2026-01-21 05:54:00 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:796779bb projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-796779bb-f2fds eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali6cfc639d9e3 [] [] }} ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-f2fds" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--f2fds-" Jan 21 05:56:32.581620 containerd[1639]: 2026-01-21 05:56:31.474 [INFO][5396] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-f2fds" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" Jan 21 05:56:32.581620 containerd[1639]: 2026-01-21 05:56:31.907 [INFO][5428] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" HandleID="k8s-pod-network.52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Workload="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:31.908 [INFO][5428] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" HandleID="k8s-pod-network.52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Workload="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00038dbb0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-796779bb-f2fds", "timestamp":"2026-01-21 05:56:31.90717551 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:31.908 [INFO][5428] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:31.908 [INFO][5428] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:31.908 [INFO][5428] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:31.945 [INFO][5428] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" host="localhost" Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:32.065 [INFO][5428] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:32.142 [INFO][5428] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:32.159 [INFO][5428] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:32.223 [INFO][5428] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:32.582203 containerd[1639]: 2026-01-21 05:56:32.224 [INFO][5428] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" host="localhost" Jan 21 05:56:32.587503 containerd[1639]: 2026-01-21 05:56:32.235 [INFO][5428] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669 Jan 21 05:56:32.587503 containerd[1639]: 2026-01-21 05:56:32.262 [INFO][5428] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" host="localhost" Jan 21 05:56:32.587503 containerd[1639]: 2026-01-21 05:56:32.332 [INFO][5428] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" host="localhost" Jan 21 05:56:32.587503 containerd[1639]: 2026-01-21 05:56:32.336 [INFO][5428] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" host="localhost" Jan 21 05:56:32.587503 containerd[1639]: 2026-01-21 05:56:32.338 [INFO][5428] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 21 05:56:32.587503 containerd[1639]: 2026-01-21 05:56:32.339 [INFO][5428] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" HandleID="k8s-pod-network.52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Workload="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" Jan 21 05:56:32.587874 containerd[1639]: 2026-01-21 05:56:32.400 [INFO][5396] cni-plugin/k8s.go 418: Populated endpoint ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-f2fds" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--796779bb--f2fds-eth0", GenerateName:"calico-apiserver-796779bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"dac6efad-1039-4242-b1c8-782ac61e5470", ResourceVersion:"1091", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"796779bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-796779bb-f2fds", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6cfc639d9e3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:32.588134 containerd[1639]: 2026-01-21 05:56:32.413 [INFO][5396] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-f2fds" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" Jan 21 05:56:32.588134 containerd[1639]: 2026-01-21 05:56:32.413 [INFO][5396] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6cfc639d9e3 ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-f2fds" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" Jan 21 05:56:32.588134 containerd[1639]: 2026-01-21 05:56:32.495 [INFO][5396] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-f2fds" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" Jan 21 05:56:32.594949 containerd[1639]: 2026-01-21 05:56:32.502 [INFO][5396] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-f2fds" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--796779bb--f2fds-eth0", GenerateName:"calico-apiserver-796779bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"dac6efad-1039-4242-b1c8-782ac61e5470", ResourceVersion:"1091", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"796779bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669", Pod:"calico-apiserver-796779bb-f2fds", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6cfc639d9e3", MAC:"46:28:e1:86:fe:74", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:32.595165 containerd[1639]: 2026-01-21 05:56:32.567 [INFO][5396] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-f2fds" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--f2fds-eth0" Jan 21 05:56:32.706000 audit: BPF prog-id=173 op=LOAD Jan 21 05:56:32.710000 audit: BPF prog-id=174 op=LOAD Jan 21 05:56:32.710000 audit[5559]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=5540 pid=5559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:32.710000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3438353865333333376563373362323039313931313830663730336133 Jan 21 05:56:32.712000 audit: BPF prog-id=174 op=UNLOAD Jan 21 05:56:32.712000 audit[5559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5540 pid=5559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:32.712000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3438353865333333376563373362323039313931313830663730336133 Jan 21 05:56:32.714000 audit: BPF prog-id=175 op=LOAD Jan 21 05:56:32.714000 audit[5559]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=5540 pid=5559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:32.714000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3438353865333333376563373362323039313931313830663730336133 Jan 21 05:56:32.715000 audit: BPF prog-id=176 op=LOAD Jan 21 05:56:32.715000 audit[5559]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=5540 pid=5559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:32.715000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3438353865333333376563373362323039313931313830663730336133 Jan 21 05:56:32.717000 audit: BPF prog-id=176 op=UNLOAD Jan 21 05:56:32.717000 audit[5559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5540 pid=5559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:32.717000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3438353865333333376563373362323039313931313830663730336133 Jan 21 05:56:32.717000 audit: BPF prog-id=175 op=UNLOAD Jan 21 05:56:32.717000 audit[5559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5540 pid=5559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:32.717000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3438353865333333376563373362323039313931313830663730336133 Jan 21 05:56:32.717000 audit: BPF prog-id=177 op=LOAD Jan 21 05:56:32.717000 audit[5559]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=5540 pid=5559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:32.717000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3438353865333333376563373362323039313931313830663730336133 Jan 21 05:56:32.734453 systemd-resolved[1290]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 21 05:56:32.752996 containerd[1639]: time="2026-01-21T05:56:32.751970116Z" level=info msg="connecting to shim 52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669" address="unix:///run/containerd/s/262d485b9b8b958a8cf2adf5d0462551133cdfe40b8c45b25f2b7b34dad9b988" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:56:32.948450 sshd[5551]: Connection closed by 10.0.0.1 port 42622 Jan 21 05:56:32.954789 sshd-session[5447]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:32.961000 audit[5447]: USER_END pid=5447 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:32.962000 audit[5447]: CRED_DISP pid=5447 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:32.980332 systemd-networkd[1528]: cali3b001fe5b05: Gained IPv6LL Jan 21 05:56:32.982036 systemd[1]: sshd@16-10.0.0.61:22-10.0.0.1:42622.service: Deactivated successfully. Jan 21 05:56:32.983000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.61:22-10.0.0.1:42622 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:32.996470 systemd[1]: session-18.scope: Deactivated successfully. Jan 21 05:56:33.022000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.61:22-10.0.0.1:42626 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:33.022957 systemd[1]: Started sshd@17-10.0.0.61:22-10.0.0.1:42626.service - OpenSSH per-connection server daemon (10.0.0.1:42626). Jan 21 05:56:33.029371 systemd-logind[1620]: Session 18 logged out. Waiting for processes to exit. Jan 21 05:56:33.044600 systemd-logind[1620]: Removed session 18. Jan 21 05:56:33.090019 systemd[1]: Started cri-containerd-52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669.scope - libcontainer container 52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669. Jan 21 05:56:33.235181 kubelet[2982]: E0121 05:56:33.234556 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:33.239189 containerd[1639]: time="2026-01-21T05:56:33.239122972Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:56:33.240832 containerd[1639]: time="2026-01-21T05:56:33.240024247Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,}" Jan 21 05:56:33.240832 containerd[1639]: time="2026-01-21T05:56:33.240101743Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,}" Jan 21 05:56:33.283330 containerd[1639]: time="2026-01-21T05:56:33.282997618Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-cd95c566-k55c2,Uid:da0d7215-59d8-4e4a-9787-c2942086bde3,Namespace:calico-system,Attempt:0,} returns sandbox id \"4858e3337ec73b209191180f703a35ac2479199fe95ffdcdb3c1f309388a762a\"" Jan 21 05:56:33.294363 containerd[1639]: time="2026-01-21T05:56:33.294223608Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 21 05:56:33.324000 audit[5630]: USER_ACCT pid=5630 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:33.326340 sshd[5630]: Accepted publickey for core from 10.0.0.1 port 42626 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:33.330000 audit[5630]: CRED_ACQ pid=5630 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:33.330000 audit[5630]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe187646a0 a2=3 a3=0 items=0 ppid=1 pid=5630 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:33.330000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:33.334563 sshd-session[5630]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:33.363546 systemd-logind[1620]: New session 19 of user core. Jan 21 05:56:33.380594 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 21 05:56:33.407000 audit[5630]: USER_START pid=5630 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:33.425000 audit[5660]: CRED_ACQ pid=5660 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:33.456000 audit: BPF prog-id=178 op=LOAD Jan 21 05:56:33.461000 audit: BPF prog-id=179 op=LOAD Jan 21 05:56:33.461000 audit[5614]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000174238 a2=98 a3=0 items=0 ppid=5598 pid=5614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:33.461000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3532616566393539656266373237616537353063393361616362346561 Jan 21 05:56:33.462000 audit: BPF prog-id=179 op=UNLOAD Jan 21 05:56:33.462000 audit[5614]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5598 pid=5614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:33.462000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3532616566393539656266373237616537353063393361616362346561 Jan 21 05:56:33.471000 audit: BPF prog-id=180 op=LOAD Jan 21 05:56:33.471000 audit[5614]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000174488 a2=98 a3=0 items=0 ppid=5598 pid=5614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:33.471000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3532616566393539656266373237616537353063393361616362346561 Jan 21 05:56:33.477000 audit: BPF prog-id=181 op=LOAD Jan 21 05:56:33.477000 audit[5614]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000174218 a2=98 a3=0 items=0 ppid=5598 pid=5614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:33.477000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3532616566393539656266373237616537353063393361616362346561 Jan 21 05:56:33.477000 audit: BPF prog-id=181 op=UNLOAD Jan 21 05:56:33.477000 audit[5614]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=5598 pid=5614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:33.477000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3532616566393539656266373237616537353063393361616362346561 Jan 21 05:56:33.477000 audit: BPF prog-id=180 op=UNLOAD Jan 21 05:56:33.477000 audit[5614]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5598 pid=5614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:33.477000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3532616566393539656266373237616537353063393361616362346561 Jan 21 05:56:33.510000 audit: BPF prog-id=182 op=LOAD Jan 21 05:56:33.510000 audit[5614]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001746e8 a2=98 a3=0 items=0 ppid=5598 pid=5614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:33.510000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3532616566393539656266373237616537353063393361616362346561 Jan 21 05:56:33.524540 systemd-resolved[1290]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 21 05:56:33.712824 containerd[1639]: time="2026-01-21T05:56:33.692994584Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:33.757094 containerd[1639]: time="2026-01-21T05:56:33.755869899Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 21 05:56:33.757094 containerd[1639]: time="2026-01-21T05:56:33.756173747Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:33.758589 kubelet[2982]: E0121 05:56:33.757329 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 21 05:56:33.758589 kubelet[2982]: E0121 05:56:33.757459 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 21 05:56:33.761374 kubelet[2982]: E0121 05:56:33.761225 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:2d409fc375e643af9c78cec77dff7f0b,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rgkt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-cd95c566-k55c2_calico-system(da0d7215-59d8-4e4a-9787-c2942086bde3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:33.775956 containerd[1639]: time="2026-01-21T05:56:33.775907997Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 21 05:56:33.947380 systemd-networkd[1528]: cali6cfc639d9e3: Gained IPv6LL Jan 21 05:56:33.949037 sshd[5660]: Connection closed by 10.0.0.1 port 42626 Jan 21 05:56:33.949793 sshd-session[5630]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:33.959000 audit[5630]: USER_END pid=5630 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:33.960000 audit[5630]: CRED_DISP pid=5630 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:33.973000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.61:22-10.0.0.1:42626 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:33.973443 systemd[1]: sshd@17-10.0.0.61:22-10.0.0.1:42626.service: Deactivated successfully. Jan 21 05:56:33.977047 systemd-logind[1620]: Session 19 logged out. Waiting for processes to exit. Jan 21 05:56:33.984813 systemd[1]: session-19.scope: Deactivated successfully. Jan 21 05:56:33.998544 systemd-logind[1620]: Removed session 19. Jan 21 05:56:34.007966 containerd[1639]: time="2026-01-21T05:56:34.007891150Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:34.054798 containerd[1639]: time="2026-01-21T05:56:34.052623818Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 21 05:56:34.054798 containerd[1639]: time="2026-01-21T05:56:34.052925883Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:34.059529 kubelet[2982]: E0121 05:56:34.057833 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 21 05:56:34.059529 kubelet[2982]: E0121 05:56:34.057912 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 21 05:56:34.062414 kubelet[2982]: E0121 05:56:34.061989 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rgkt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-cd95c566-k55c2_calico-system(da0d7215-59d8-4e4a-9787-c2942086bde3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:34.065519 kubelet[2982]: E0121 05:56:34.065085 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-cd95c566-k55c2" podUID="da0d7215-59d8-4e4a-9787-c2942086bde3" Jan 21 05:56:34.220000 audit: BPF prog-id=183 op=LOAD Jan 21 05:56:34.220000 audit[5748]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe7ab2bce0 a2=98 a3=1fffffffffffffff items=0 ppid=5422 pid=5748 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.220000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 21 05:56:34.220000 audit: BPF prog-id=183 op=UNLOAD Jan 21 05:56:34.220000 audit[5748]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffe7ab2bcb0 a3=0 items=0 ppid=5422 pid=5748 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.220000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 21 05:56:34.221000 audit: BPF prog-id=184 op=LOAD Jan 21 05:56:34.221000 audit[5748]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe7ab2bbc0 a2=94 a3=3 items=0 ppid=5422 pid=5748 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.221000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 21 05:56:34.221000 audit: BPF prog-id=184 op=UNLOAD Jan 21 05:56:34.221000 audit[5748]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffe7ab2bbc0 a2=94 a3=3 items=0 ppid=5422 pid=5748 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.221000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 21 05:56:34.221000 audit: BPF prog-id=185 op=LOAD Jan 21 05:56:34.221000 audit[5748]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe7ab2bc00 a2=94 a3=7ffe7ab2bde0 items=0 ppid=5422 pid=5748 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.221000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 21 05:56:34.221000 audit: BPF prog-id=185 op=UNLOAD Jan 21 05:56:34.221000 audit[5748]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffe7ab2bc00 a2=94 a3=7ffe7ab2bde0 items=0 ppid=5422 pid=5748 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.221000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 21 05:56:34.234822 kubelet[2982]: E0121 05:56:34.215540 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:34.234939 containerd[1639]: time="2026-01-21T05:56:34.231353051Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,}" Jan 21 05:56:34.244837 containerd[1639]: time="2026-01-21T05:56:34.243234910Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,}" Jan 21 05:56:34.309000 audit: BPF prog-id=186 op=LOAD Jan 21 05:56:34.309000 audit[5756]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe945ca940 a2=98 a3=3 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.309000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:34.316000 audit: BPF prog-id=186 op=UNLOAD Jan 21 05:56:34.316000 audit[5756]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffe945ca910 a3=0 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.316000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:34.324000 audit: BPF prog-id=187 op=LOAD Jan 21 05:56:34.324000 audit[5756]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe945ca730 a2=94 a3=54428f items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.324000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:34.324000 audit: BPF prog-id=187 op=UNLOAD Jan 21 05:56:34.324000 audit[5756]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe945ca730 a2=94 a3=54428f items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.324000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:34.324000 audit: BPF prog-id=188 op=LOAD Jan 21 05:56:34.324000 audit[5756]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe945ca760 a2=94 a3=2 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.324000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:34.331000 audit: BPF prog-id=188 op=UNLOAD Jan 21 05:56:34.331000 audit[5756]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe945ca760 a2=0 a3=2 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:34.331000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:34.710735 containerd[1639]: time="2026-01-21T05:56:34.707554550Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-f2fds,Uid:dac6efad-1039-4242-b1c8-782ac61e5470,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"52aef959ebf727ae750c93aacb4ea2adc8c149da2548e372ed6cff539095b669\"" Jan 21 05:56:34.734077 containerd[1639]: time="2026-01-21T05:56:34.732920661Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 21 05:56:35.208500 kubelet[2982]: E0121 05:56:35.196471 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-cd95c566-k55c2" podUID="da0d7215-59d8-4e4a-9787-c2942086bde3" Jan 21 05:56:35.352827 containerd[1639]: time="2026-01-21T05:56:35.352457382Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:35.359622 containerd[1639]: time="2026-01-21T05:56:35.359135025Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 21 05:56:35.359622 containerd[1639]: time="2026-01-21T05:56:35.359367880Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:35.361337 kubelet[2982]: E0121 05:56:35.361194 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:35.361448 kubelet[2982]: E0121 05:56:35.361362 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:35.361858 kubelet[2982]: E0121 05:56:35.361551 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n5wjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:35.361858 kubelet[2982]: E0121 05:56:35.362840 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:56:35.607000 audit[5805]: NETFILTER_CFG table=filter:121 family=2 entries=20 op=nft_register_rule pid=5805 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:35.607000 audit[5805]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc2b4d04a0 a2=0 a3=7ffc2b4d048c items=0 ppid=3144 pid=5805 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:35.607000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:35.621000 audit[5805]: NETFILTER_CFG table=nat:122 family=2 entries=14 op=nft_register_rule pid=5805 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:35.621000 audit[5805]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffc2b4d04a0 a2=0 a3=0 items=0 ppid=3144 pid=5805 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:35.621000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:35.747849 systemd-networkd[1528]: cali7b90fb0b84a: Link UP Jan 21 05:56:35.752372 systemd-networkd[1528]: cali7b90fb0b84a: Gained carrier Jan 21 05:56:35.824631 kubelet[2982]: E0121 05:56:35.824403 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:56:36.024176 kernel: kauditd_printk_skb: 109 callbacks suppressed Jan 21 05:56:36.024437 kernel: audit: type=1325 audit(1768974995.999:691): table=filter:123 family=2 entries=20 op=nft_register_rule pid=5826 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:35.999000 audit[5826]: NETFILTER_CFG table=filter:123 family=2 entries=20 op=nft_register_rule pid=5826 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:36.025345 containerd[1639]: 2026-01-21 05:56:34.151 [INFO][5651] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0 calico-apiserver-fcddb958d- calico-apiserver dbddb9a5-130c-465d-9677-8a7b919a18f8 1096 0 2026-01-21 05:54:01 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:fcddb958d projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-fcddb958d-sjlxh eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali7b90fb0b84a [] [] }} ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Namespace="calico-apiserver" Pod="calico-apiserver-fcddb958d-sjlxh" WorkloadEndpoint="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-" Jan 21 05:56:36.025345 containerd[1639]: 2026-01-21 05:56:34.155 [INFO][5651] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Namespace="calico-apiserver" Pod="calico-apiserver-fcddb958d-sjlxh" WorkloadEndpoint="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" Jan 21 05:56:36.025345 containerd[1639]: 2026-01-21 05:56:35.380 [INFO][5744] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" HandleID="k8s-pod-network.8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Workload="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.385 [INFO][5744] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" HandleID="k8s-pod-network.8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Workload="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0005dc3a0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-fcddb958d-sjlxh", "timestamp":"2026-01-21 05:56:35.380611913 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.389 [INFO][5744] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.389 [INFO][5744] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.389 [INFO][5744] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.438 [INFO][5744] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" host="localhost" Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.502 [INFO][5744] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.534 [INFO][5744] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.570 [INFO][5744] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.585 [INFO][5744] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:36.027372 containerd[1639]: 2026-01-21 05:56:35.589 [INFO][5744] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" host="localhost" Jan 21 05:56:36.029044 containerd[1639]: 2026-01-21 05:56:35.604 [INFO][5744] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d Jan 21 05:56:36.029044 containerd[1639]: 2026-01-21 05:56:35.663 [INFO][5744] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" host="localhost" Jan 21 05:56:36.029044 containerd[1639]: 2026-01-21 05:56:35.691 [INFO][5744] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" host="localhost" Jan 21 05:56:36.029044 containerd[1639]: 2026-01-21 05:56:35.693 [INFO][5744] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" host="localhost" Jan 21 05:56:36.029044 containerd[1639]: 2026-01-21 05:56:35.694 [INFO][5744] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 21 05:56:36.029044 containerd[1639]: 2026-01-21 05:56:35.694 [INFO][5744] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" HandleID="k8s-pod-network.8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Workload="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" Jan 21 05:56:36.029328 containerd[1639]: 2026-01-21 05:56:35.713 [INFO][5651] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Namespace="calico-apiserver" Pod="calico-apiserver-fcddb958d-sjlxh" WorkloadEndpoint="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0", GenerateName:"calico-apiserver-fcddb958d-", Namespace:"calico-apiserver", SelfLink:"", UID:"dbddb9a5-130c-465d-9677-8a7b919a18f8", ResourceVersion:"1096", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fcddb958d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-fcddb958d-sjlxh", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali7b90fb0b84a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:36.029529 containerd[1639]: 2026-01-21 05:56:35.713 [INFO][5651] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Namespace="calico-apiserver" Pod="calico-apiserver-fcddb958d-sjlxh" WorkloadEndpoint="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" Jan 21 05:56:36.029529 containerd[1639]: 2026-01-21 05:56:35.713 [INFO][5651] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7b90fb0b84a ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Namespace="calico-apiserver" Pod="calico-apiserver-fcddb958d-sjlxh" WorkloadEndpoint="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" Jan 21 05:56:36.029529 containerd[1639]: 2026-01-21 05:56:35.807 [INFO][5651] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Namespace="calico-apiserver" Pod="calico-apiserver-fcddb958d-sjlxh" WorkloadEndpoint="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" Jan 21 05:56:36.032831 containerd[1639]: 2026-01-21 05:56:35.856 [INFO][5651] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Namespace="calico-apiserver" Pod="calico-apiserver-fcddb958d-sjlxh" WorkloadEndpoint="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0", GenerateName:"calico-apiserver-fcddb958d-", Namespace:"calico-apiserver", SelfLink:"", UID:"dbddb9a5-130c-465d-9677-8a7b919a18f8", ResourceVersion:"1096", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fcddb958d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d", Pod:"calico-apiserver-fcddb958d-sjlxh", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali7b90fb0b84a", MAC:"2e:7c:b0:52:38:1a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:35.999000 audit[5826]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffce28a4f10 a2=0 a3=7ffce28a4efc items=0 ppid=3144 pid=5826 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.033234 containerd[1639]: 2026-01-21 05:56:35.954 [INFO][5651] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" Namespace="calico-apiserver" Pod="calico-apiserver-fcddb958d-sjlxh" WorkloadEndpoint="localhost-k8s-calico--apiserver--fcddb958d--sjlxh-eth0" Jan 21 05:56:36.069075 kernel: audit: type=1300 audit(1768974995.999:691): arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffce28a4f10 a2=0 a3=7ffce28a4efc items=0 ppid=3144 pid=5826 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:35.999000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:36.031000 audit[5826]: NETFILTER_CFG table=nat:124 family=2 entries=14 op=nft_register_rule pid=5826 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:36.111012 kernel: audit: type=1327 audit(1768974995.999:691): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:36.112036 kernel: audit: type=1325 audit(1768974996.031:692): table=nat:124 family=2 entries=14 op=nft_register_rule pid=5826 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:36.112088 kernel: audit: type=1300 audit(1768974996.031:692): arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffce28a4f10 a2=0 a3=0 items=0 ppid=3144 pid=5826 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.031000 audit[5826]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffce28a4f10 a2=0 a3=0 items=0 ppid=3144 pid=5826 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.170778 kernel: audit: type=1327 audit(1768974996.031:692): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:36.031000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:36.232608 systemd-networkd[1528]: cali73f5a5709c6: Link UP Jan 21 05:56:36.247594 systemd-networkd[1528]: cali73f5a5709c6: Gained carrier Jan 21 05:56:36.305000 audit: BPF prog-id=189 op=LOAD Jan 21 05:56:36.393537 kernel: audit: type=1334 audit(1768974996.305:693): prog-id=189 op=LOAD Jan 21 05:56:36.393780 kernel: audit: type=1300 audit(1768974996.305:693): arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe945ca620 a2=94 a3=1 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.305000 audit[5756]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe945ca620 a2=94 a3=1 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.403622 containerd[1639]: time="2026-01-21T05:56:36.318505599Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,}" Jan 21 05:56:36.419099 kernel: audit: type=1327 audit(1768974996.305:693): proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.305000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.462035 kernel: audit: type=1334 audit(1768974996.305:694): prog-id=189 op=UNLOAD Jan 21 05:56:36.305000 audit: BPF prog-id=189 op=UNLOAD Jan 21 05:56:36.305000 audit[5756]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe945ca620 a2=94 a3=1 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.305000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.472000 audit: BPF prog-id=190 op=LOAD Jan 21 05:56:36.472000 audit[5756]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe945ca610 a2=94 a3=4 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.472000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.474000 audit: BPF prog-id=190 op=UNLOAD Jan 21 05:56:36.474000 audit[5756]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe945ca610 a2=0 a3=4 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.474000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.479000 audit: BPF prog-id=191 op=LOAD Jan 21 05:56:36.479000 audit[5756]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffe945ca470 a2=94 a3=5 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.479000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.479000 audit: BPF prog-id=191 op=UNLOAD Jan 21 05:56:36.479000 audit[5756]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffe945ca470 a2=0 a3=5 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.479000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.485000 audit: BPF prog-id=192 op=LOAD Jan 21 05:56:36.485000 audit[5756]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe945ca690 a2=94 a3=6 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.485000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.486000 audit: BPF prog-id=192 op=UNLOAD Jan 21 05:56:36.486000 audit[5756]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe945ca690 a2=0 a3=6 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.486000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.488000 audit: BPF prog-id=193 op=LOAD Jan 21 05:56:36.488000 audit[5756]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe945c9e40 a2=94 a3=88 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.488000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.497000 audit: BPF prog-id=194 op=LOAD Jan 21 05:56:36.497000 audit[5756]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7ffe945c9cc0 a2=94 a3=2 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.497000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.497000 audit: BPF prog-id=194 op=UNLOAD Jan 21 05:56:36.497000 audit[5756]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7ffe945c9cf0 a2=0 a3=7ffe945c9df0 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.497000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.514152 containerd[1639]: time="2026-01-21T05:56:36.513794322Z" level=info msg="connecting to shim 8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d" address="unix:///run/containerd/s/f4bcab24757d7c64bc52673f964be07c6c6c33b4a3cd4d4d9820775755bc3d36" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:56:36.513000 audit: BPF prog-id=193 op=UNLOAD Jan 21 05:56:36.513000 audit[5756]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=3cb94d10 a2=0 a3=4d1587efb69ce376 items=0 ppid=5422 pid=5756 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.513000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 21 05:56:36.525781 containerd[1639]: 2026-01-21 05:56:34.287 [INFO][5674] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0 coredns-674b8bbfcf- kube-system 36de057e-0cdd-42d8-a113-435fdf1d98e9 1101 0 2026-01-21 05:53:29 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-sgzh7 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali73f5a5709c6 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Namespace="kube-system" Pod="coredns-674b8bbfcf-sgzh7" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--sgzh7-" Jan 21 05:56:36.525781 containerd[1639]: 2026-01-21 05:56:34.288 [INFO][5674] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Namespace="kube-system" Pod="coredns-674b8bbfcf-sgzh7" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" Jan 21 05:56:36.525781 containerd[1639]: 2026-01-21 05:56:35.472 [INFO][5765] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" HandleID="k8s-pod-network.9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Workload="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.516 [INFO][5765] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" HandleID="k8s-pod-network.9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Workload="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000515c00), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-sgzh7", "timestamp":"2026-01-21 05:56:35.472153365 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.516 [INFO][5765] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.715 [INFO][5765] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.716 [INFO][5765] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.762 [INFO][5765] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" host="localhost" Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.790 [INFO][5765] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.870 [INFO][5765] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.948 [INFO][5765] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.987 [INFO][5765] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:36.526136 containerd[1639]: 2026-01-21 05:56:35.987 [INFO][5765] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" host="localhost" Jan 21 05:56:36.533187 containerd[1639]: 2026-01-21 05:56:36.025 [INFO][5765] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7 Jan 21 05:56:36.533187 containerd[1639]: 2026-01-21 05:56:36.068 [INFO][5765] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" host="localhost" Jan 21 05:56:36.533187 containerd[1639]: 2026-01-21 05:56:36.125 [INFO][5765] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" host="localhost" Jan 21 05:56:36.533187 containerd[1639]: 2026-01-21 05:56:36.130 [INFO][5765] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" host="localhost" Jan 21 05:56:36.533187 containerd[1639]: 2026-01-21 05:56:36.130 [INFO][5765] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 21 05:56:36.533187 containerd[1639]: 2026-01-21 05:56:36.130 [INFO][5765] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" HandleID="k8s-pod-network.9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Workload="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" Jan 21 05:56:36.533488 containerd[1639]: 2026-01-21 05:56:36.181 [INFO][5674] cni-plugin/k8s.go 418: Populated endpoint ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Namespace="kube-system" Pod="coredns-674b8bbfcf-sgzh7" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"36de057e-0cdd-42d8-a113-435fdf1d98e9", ResourceVersion:"1101", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 53, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-sgzh7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali73f5a5709c6", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:36.533810 containerd[1639]: 2026-01-21 05:56:36.181 [INFO][5674] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Namespace="kube-system" Pod="coredns-674b8bbfcf-sgzh7" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" Jan 21 05:56:36.533810 containerd[1639]: 2026-01-21 05:56:36.181 [INFO][5674] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali73f5a5709c6 ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Namespace="kube-system" Pod="coredns-674b8bbfcf-sgzh7" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" Jan 21 05:56:36.533810 containerd[1639]: 2026-01-21 05:56:36.254 [INFO][5674] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Namespace="kube-system" Pod="coredns-674b8bbfcf-sgzh7" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" Jan 21 05:56:36.533915 containerd[1639]: 2026-01-21 05:56:36.256 [INFO][5674] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Namespace="kube-system" Pod="coredns-674b8bbfcf-sgzh7" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"36de057e-0cdd-42d8-a113-435fdf1d98e9", ResourceVersion:"1101", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 53, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7", Pod:"coredns-674b8bbfcf-sgzh7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali73f5a5709c6", MAC:"a2:35:64:eb:a0:d6", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:36.533915 containerd[1639]: 2026-01-21 05:56:36.479 [INFO][5674] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" Namespace="kube-system" Pod="coredns-674b8bbfcf-sgzh7" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--sgzh7-eth0" Jan 21 05:56:36.696000 audit: BPF prog-id=195 op=LOAD Jan 21 05:56:36.696000 audit[5877]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd97faa4d0 a2=98 a3=1999999999999999 items=0 ppid=5422 pid=5877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.696000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 21 05:56:36.696000 audit: BPF prog-id=195 op=UNLOAD Jan 21 05:56:36.696000 audit[5877]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffd97faa4a0 a3=0 items=0 ppid=5422 pid=5877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.696000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 21 05:56:36.697000 audit: BPF prog-id=196 op=LOAD Jan 21 05:56:36.697000 audit[5877]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd97faa3b0 a2=94 a3=ffff items=0 ppid=5422 pid=5877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.697000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 21 05:56:36.697000 audit: BPF prog-id=196 op=UNLOAD Jan 21 05:56:36.697000 audit[5877]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd97faa3b0 a2=94 a3=ffff items=0 ppid=5422 pid=5877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.697000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 21 05:56:36.697000 audit: BPF prog-id=197 op=LOAD Jan 21 05:56:36.697000 audit[5877]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd97faa3f0 a2=94 a3=7ffd97faa5d0 items=0 ppid=5422 pid=5877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.697000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 21 05:56:36.697000 audit: BPF prog-id=197 op=UNLOAD Jan 21 05:56:36.697000 audit[5877]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd97faa3f0 a2=94 a3=7ffd97faa5d0 items=0 ppid=5422 pid=5877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:36.697000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 21 05:56:36.800632 systemd[1]: Started cri-containerd-8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d.scope - libcontainer container 8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d. Jan 21 05:56:36.874841 containerd[1639]: time="2026-01-21T05:56:36.874536025Z" level=info msg="connecting to shim 9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7" address="unix:///run/containerd/s/f8187a8e8e6b66fda912785ed062d55da4a4b3547deb157ace276fddd200645a" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:56:36.917449 systemd-networkd[1528]: caliac5641da4f9: Link UP Jan 21 05:56:36.919012 systemd-networkd[1528]: caliac5641da4f9: Gained carrier Jan 21 05:56:37.066000 audit: BPF prog-id=198 op=LOAD Jan 21 05:56:37.089000 audit: BPF prog-id=199 op=LOAD Jan 21 05:56:37.089000 audit[5876]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=5845 pid=5876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.089000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3862333831616164376438663435626439306135633265626565626164 Jan 21 05:56:37.091000 audit: BPF prog-id=199 op=UNLOAD Jan 21 05:56:37.091000 audit[5876]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5845 pid=5876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.091000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3862333831616164376438663435626439306135633265626565626164 Jan 21 05:56:37.102000 audit: BPF prog-id=200 op=LOAD Jan 21 05:56:37.102000 audit[5876]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=5845 pid=5876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.102000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3862333831616164376438663435626439306135633265626565626164 Jan 21 05:56:37.103000 audit: BPF prog-id=201 op=LOAD Jan 21 05:56:37.103000 audit[5876]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=5845 pid=5876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.103000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3862333831616164376438663435626439306135633265626565626164 Jan 21 05:56:37.103000 audit: BPF prog-id=201 op=UNLOAD Jan 21 05:56:37.103000 audit[5876]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5845 pid=5876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.103000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3862333831616164376438663435626439306135633265626565626164 Jan 21 05:56:37.104000 audit: BPF prog-id=200 op=UNLOAD Jan 21 05:56:37.104000 audit[5876]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5845 pid=5876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.104000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3862333831616164376438663435626439306135633265626565626164 Jan 21 05:56:37.104000 audit: BPF prog-id=202 op=LOAD Jan 21 05:56:37.104000 audit[5876]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=5845 pid=5876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.104000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3862333831616164376438663435626439306135633265626565626164 Jan 21 05:56:37.123626 systemd-resolved[1290]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:34.175 [INFO][5666] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--zftdp-eth0 csi-node-driver- calico-system 487c6aa2-04b5-4684-9420-0e4b73799fd3 892 0 2026-01-21 05:54:33 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-zftdp eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] caliac5641da4f9 [] [] }} ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Namespace="calico-system" Pod="csi-node-driver-zftdp" WorkloadEndpoint="localhost-k8s-csi--node--driver--zftdp-" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:34.176 [INFO][5666] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Namespace="calico-system" Pod="csi-node-driver-zftdp" WorkloadEndpoint="localhost-k8s-csi--node--driver--zftdp-eth0" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:35.528 [INFO][5755] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" HandleID="k8s-pod-network.28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Workload="localhost-k8s-csi--node--driver--zftdp-eth0" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:35.534 [INFO][5755] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" HandleID="k8s-pod-network.28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Workload="localhost-k8s-csi--node--driver--zftdp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000426720), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-zftdp", "timestamp":"2026-01-21 05:56:35.52892347 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:35.534 [INFO][5755] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.130 [INFO][5755] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.132 [INFO][5755] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.204 [INFO][5755] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" host="localhost" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.490 [INFO][5755] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.574 [INFO][5755] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.633 [INFO][5755] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.655 [INFO][5755] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.657 [INFO][5755] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" host="localhost" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.670 [INFO][5755] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.746 [INFO][5755] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" host="localhost" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.815 [INFO][5755] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" host="localhost" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.815 [INFO][5755] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" host="localhost" Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.816 [INFO][5755] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 21 05:56:37.150430 containerd[1639]: 2026-01-21 05:56:36.817 [INFO][5755] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" HandleID="k8s-pod-network.28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Workload="localhost-k8s-csi--node--driver--zftdp-eth0" Jan 21 05:56:37.152099 containerd[1639]: 2026-01-21 05:56:36.853 [INFO][5666] cni-plugin/k8s.go 418: Populated endpoint ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Namespace="calico-system" Pod="csi-node-driver-zftdp" WorkloadEndpoint="localhost-k8s-csi--node--driver--zftdp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--zftdp-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"487c6aa2-04b5-4684-9420-0e4b73799fd3", ResourceVersion:"892", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-zftdp", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"caliac5641da4f9", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:37.152099 containerd[1639]: 2026-01-21 05:56:36.854 [INFO][5666] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Namespace="calico-system" Pod="csi-node-driver-zftdp" WorkloadEndpoint="localhost-k8s-csi--node--driver--zftdp-eth0" Jan 21 05:56:37.152099 containerd[1639]: 2026-01-21 05:56:36.854 [INFO][5666] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliac5641da4f9 ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Namespace="calico-system" Pod="csi-node-driver-zftdp" WorkloadEndpoint="localhost-k8s-csi--node--driver--zftdp-eth0" Jan 21 05:56:37.152099 containerd[1639]: 2026-01-21 05:56:36.931 [INFO][5666] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Namespace="calico-system" Pod="csi-node-driver-zftdp" WorkloadEndpoint="localhost-k8s-csi--node--driver--zftdp-eth0" Jan 21 05:56:37.152099 containerd[1639]: 2026-01-21 05:56:36.942 [INFO][5666] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Namespace="calico-system" Pod="csi-node-driver-zftdp" WorkloadEndpoint="localhost-k8s-csi--node--driver--zftdp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--zftdp-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"487c6aa2-04b5-4684-9420-0e4b73799fd3", ResourceVersion:"892", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c", Pod:"csi-node-driver-zftdp", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"caliac5641da4f9", MAC:"e2:61:ed:70:2e:9c", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:37.152099 containerd[1639]: 2026-01-21 05:56:37.046 [INFO][5666] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" Namespace="calico-system" Pod="csi-node-driver-zftdp" WorkloadEndpoint="localhost-k8s-csi--node--driver--zftdp-eth0" Jan 21 05:56:37.437164 systemd-networkd[1528]: cali106c2b932fa: Link UP Jan 21 05:56:37.446880 systemd-networkd[1528]: cali106c2b932fa: Gained carrier Jan 21 05:56:37.466423 systemd-networkd[1528]: cali7b90fb0b84a: Gained IPv6LL Jan 21 05:56:37.552123 systemd[1]: Started cri-containerd-9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7.scope - libcontainer container 9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7. Jan 21 05:56:37.792870 systemd-networkd[1528]: vxlan.calico: Link UP Jan 21 05:56:37.793052 systemd-networkd[1528]: vxlan.calico: Gained carrier Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:35.567 [INFO][5771] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--666569f655--m7pjj-eth0 goldmane-666569f655- calico-system 59842dba-2f25-4ec6-86e9-919aa2afb77a 1098 0 2026-01-21 05:54:21 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-666569f655-m7pjj eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali106c2b932fa [] [] }} ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Namespace="calico-system" Pod="goldmane-666569f655-m7pjj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--m7pjj-" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:35.567 [INFO][5771] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Namespace="calico-system" Pod="goldmane-666569f655-m7pjj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--m7pjj-eth0" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:35.986 [INFO][5807] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" HandleID="k8s-pod-network.e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Workload="localhost-k8s-goldmane--666569f655--m7pjj-eth0" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:36.003 [INFO][5807] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" HandleID="k8s-pod-network.e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Workload="localhost-k8s-goldmane--666569f655--m7pjj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000513ce0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-666569f655-m7pjj", "timestamp":"2026-01-21 05:56:35.986447881 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:36.005 [INFO][5807] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:36.818 [INFO][5807] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:36.819 [INFO][5807] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:36.860 [INFO][5807] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" host="localhost" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:36.969 [INFO][5807] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.076 [INFO][5807] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.097 [INFO][5807] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.167 [INFO][5807] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.169 [INFO][5807] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" host="localhost" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.186 [INFO][5807] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.220 [INFO][5807] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" host="localhost" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.313 [INFO][5807] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" host="localhost" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.321 [INFO][5807] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" host="localhost" Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.322 [INFO][5807] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 21 05:56:37.832823 containerd[1639]: 2026-01-21 05:56:37.322 [INFO][5807] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" HandleID="k8s-pod-network.e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Workload="localhost-k8s-goldmane--666569f655--m7pjj-eth0" Jan 21 05:56:37.835212 containerd[1639]: 2026-01-21 05:56:37.410 [INFO][5771] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Namespace="calico-system" Pod="goldmane-666569f655-m7pjj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--m7pjj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--m7pjj-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"59842dba-2f25-4ec6-86e9-919aa2afb77a", ResourceVersion:"1098", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-666569f655-m7pjj", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali106c2b932fa", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:37.835212 containerd[1639]: 2026-01-21 05:56:37.411 [INFO][5771] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Namespace="calico-system" Pod="goldmane-666569f655-m7pjj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--m7pjj-eth0" Jan 21 05:56:37.835212 containerd[1639]: 2026-01-21 05:56:37.411 [INFO][5771] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali106c2b932fa ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Namespace="calico-system" Pod="goldmane-666569f655-m7pjj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--m7pjj-eth0" Jan 21 05:56:37.835212 containerd[1639]: 2026-01-21 05:56:37.569 [INFO][5771] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Namespace="calico-system" Pod="goldmane-666569f655-m7pjj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--m7pjj-eth0" Jan 21 05:56:37.835212 containerd[1639]: 2026-01-21 05:56:37.659 [INFO][5771] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Namespace="calico-system" Pod="goldmane-666569f655-m7pjj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--m7pjj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--m7pjj-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"59842dba-2f25-4ec6-86e9-919aa2afb77a", ResourceVersion:"1098", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c", Pod:"goldmane-666569f655-m7pjj", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali106c2b932fa", MAC:"fa:ff:9d:a9:a6:4b", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:37.835212 containerd[1639]: 2026-01-21 05:56:37.812 [INFO][5771] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" Namespace="calico-system" Pod="goldmane-666569f655-m7pjj" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--m7pjj-eth0" Jan 21 05:56:37.901157 containerd[1639]: time="2026-01-21T05:56:37.900379509Z" level=info msg="connecting to shim 28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c" address="unix:///run/containerd/s/1648b52f6acfdec36cf859befd889a55c41082da3cf5bd811abdfae0d515c5cc" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:56:37.906000 audit: BPF prog-id=203 op=LOAD Jan 21 05:56:37.909000 audit: BPF prog-id=204 op=LOAD Jan 21 05:56:37.909000 audit[5934]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=5907 pid=5934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.909000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961616362633535366434343439363137333038333332626564613761 Jan 21 05:56:37.910000 audit: BPF prog-id=204 op=UNLOAD Jan 21 05:56:37.910000 audit[5934]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5907 pid=5934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.910000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961616362633535366434343439363137333038333332626564613761 Jan 21 05:56:37.912000 audit: BPF prog-id=205 op=LOAD Jan 21 05:56:37.912000 audit[5934]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=5907 pid=5934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.912000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961616362633535366434343439363137333038333332626564613761 Jan 21 05:56:37.914000 audit: BPF prog-id=206 op=LOAD Jan 21 05:56:37.914000 audit[5934]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=5907 pid=5934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.914000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961616362633535366434343439363137333038333332626564613761 Jan 21 05:56:37.915000 audit: BPF prog-id=206 op=UNLOAD Jan 21 05:56:37.915000 audit[5934]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5907 pid=5934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.915000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961616362633535366434343439363137333038333332626564613761 Jan 21 05:56:37.915000 audit: BPF prog-id=205 op=UNLOAD Jan 21 05:56:37.915000 audit[5934]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5907 pid=5934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.915000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961616362633535366434343439363137333038333332626564613761 Jan 21 05:56:37.916000 audit: BPF prog-id=207 op=LOAD Jan 21 05:56:37.916000 audit[5934]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=5907 pid=5934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:37.916000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961616362633535366434343439363137333038333332626564613761 Jan 21 05:56:37.928823 systemd-resolved[1290]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 21 05:56:38.106190 systemd-networkd[1528]: cali73f5a5709c6: Gained IPv6LL Jan 21 05:56:38.133567 systemd-networkd[1528]: cali402e0971676: Link UP Jan 21 05:56:38.134877 systemd-networkd[1528]: cali402e0971676: Gained carrier Jan 21 05:56:38.249820 containerd[1639]: time="2026-01-21T05:56:38.238075886Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,}" Jan 21 05:56:38.292057 containerd[1639]: time="2026-01-21T05:56:38.292000100Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fcddb958d-sjlxh,Uid:dbddb9a5-130c-465d-9677-8a7b919a18f8,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"8b381aad7d8f45bd90a5c2ebeebad63b5a30cd4df4871a41f7faea1934323a5d\"" Jan 21 05:56:38.368267 containerd[1639]: time="2026-01-21T05:56:38.367979963Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:35.563 [INFO][5761] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0 coredns-674b8bbfcf- kube-system 9bc2381f-9591-401f-8403-a07aa98276a8 1110 0 2026-01-21 05:53:29 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-qpdqz eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali402e0971676 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Namespace="kube-system" Pod="coredns-674b8bbfcf-qpdqz" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--qpdqz-" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:35.581 [INFO][5761] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Namespace="kube-system" Pod="coredns-674b8bbfcf-qpdqz" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:36.104 [INFO][5813] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" HandleID="k8s-pod-network.01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Workload="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:36.104 [INFO][5813] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" HandleID="k8s-pod-network.01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Workload="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004ff10), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-qpdqz", "timestamp":"2026-01-21 05:56:36.104010754 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:36.104 [INFO][5813] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.321 [INFO][5813] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.323 [INFO][5813] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.495 [INFO][5813] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" host="localhost" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.536 [INFO][5813] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.627 [INFO][5813] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.674 [INFO][5813] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.722 [INFO][5813] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.722 [INFO][5813] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" host="localhost" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.827 [INFO][5813] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661 Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:37.934 [INFO][5813] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" host="localhost" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:38.013 [INFO][5813] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" host="localhost" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:38.014 [INFO][5813] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" host="localhost" Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:38.016 [INFO][5813] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 21 05:56:38.379764 containerd[1639]: 2026-01-21 05:56:38.017 [INFO][5813] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" HandleID="k8s-pod-network.01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Workload="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" Jan 21 05:56:38.381059 containerd[1639]: 2026-01-21 05:56:38.045 [INFO][5761] cni-plugin/k8s.go 418: Populated endpoint ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Namespace="kube-system" Pod="coredns-674b8bbfcf-qpdqz" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"9bc2381f-9591-401f-8403-a07aa98276a8", ResourceVersion:"1110", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 53, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-qpdqz", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali402e0971676", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:38.381059 containerd[1639]: 2026-01-21 05:56:38.045 [INFO][5761] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Namespace="kube-system" Pod="coredns-674b8bbfcf-qpdqz" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" Jan 21 05:56:38.381059 containerd[1639]: 2026-01-21 05:56:38.061 [INFO][5761] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali402e0971676 ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Namespace="kube-system" Pod="coredns-674b8bbfcf-qpdqz" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" Jan 21 05:56:38.381059 containerd[1639]: 2026-01-21 05:56:38.128 [INFO][5761] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Namespace="kube-system" Pod="coredns-674b8bbfcf-qpdqz" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" Jan 21 05:56:38.381059 containerd[1639]: 2026-01-21 05:56:38.134 [INFO][5761] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Namespace="kube-system" Pod="coredns-674b8bbfcf-qpdqz" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"9bc2381f-9591-401f-8403-a07aa98276a8", ResourceVersion:"1110", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 53, 29, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661", Pod:"coredns-674b8bbfcf-qpdqz", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali402e0971676", MAC:"76:83:c6:b0:88:1f", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:38.381059 containerd[1639]: 2026-01-21 05:56:38.280 [INFO][5761] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" Namespace="kube-system" Pod="coredns-674b8bbfcf-qpdqz" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--qpdqz-eth0" Jan 21 05:56:38.455405 systemd[1]: Started cri-containerd-28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c.scope - libcontainer container 28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c. Jan 21 05:56:38.664502 containerd[1639]: time="2026-01-21T05:56:38.647181174Z" level=info msg="connecting to shim e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c" address="unix:///run/containerd/s/9c8691373a73603fa31ff36e3236b9290bb06c71f3add2beabd2f0a4aca4480c" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:56:38.672000 audit: BPF prog-id=208 op=LOAD Jan 21 05:56:38.675000 audit: BPF prog-id=209 op=LOAD Jan 21 05:56:38.675000 audit[6003]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=5978 pid=6003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:38.675000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3238613065396266613162623438653561353930663938666261383436 Jan 21 05:56:38.675000 audit: BPF prog-id=209 op=UNLOAD Jan 21 05:56:38.675000 audit[6003]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5978 pid=6003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:38.675000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3238613065396266613162623438653561353930663938666261383436 Jan 21 05:56:38.676000 audit: BPF prog-id=210 op=LOAD Jan 21 05:56:38.676000 audit[6003]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=5978 pid=6003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:38.676000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3238613065396266613162623438653561353930663938666261383436 Jan 21 05:56:38.676000 audit: BPF prog-id=211 op=LOAD Jan 21 05:56:38.676000 audit[6003]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=5978 pid=6003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:38.676000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3238613065396266613162623438653561353930663938666261383436 Jan 21 05:56:38.679000 audit: BPF prog-id=211 op=UNLOAD Jan 21 05:56:38.679000 audit[6003]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5978 pid=6003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:38.679000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3238613065396266613162623438653561353930663938666261383436 Jan 21 05:56:38.679000 audit: BPF prog-id=210 op=UNLOAD Jan 21 05:56:38.679000 audit[6003]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5978 pid=6003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:38.679000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3238613065396266613162623438653561353930663938666261383436 Jan 21 05:56:38.679000 audit: BPF prog-id=212 op=LOAD Jan 21 05:56:38.679000 audit[6003]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=5978 pid=6003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:38.679000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3238613065396266613162623438653561353930663938666261383436 Jan 21 05:56:38.682419 systemd-networkd[1528]: caliac5641da4f9: Gained IPv6LL Jan 21 05:56:38.751461 systemd-resolved[1290]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 21 05:56:38.805957 containerd[1639]: time="2026-01-21T05:56:38.801927414Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-sgzh7,Uid:36de057e-0cdd-42d8-a113-435fdf1d98e9,Namespace:kube-system,Attempt:0,} returns sandbox id \"9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7\"" Jan 21 05:56:38.819794 kubelet[2982]: E0121 05:56:38.815539 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:39.016817 containerd[1639]: time="2026-01-21T05:56:39.016763106Z" level=info msg="CreateContainer within sandbox \"9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 21 05:56:39.023266 systemd[1]: Started cri-containerd-e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c.scope - libcontainer container e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c. Jan 21 05:56:39.035000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.61:22-10.0.0.1:41746 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:39.035818 systemd[1]: Started sshd@18-10.0.0.61:22-10.0.0.1:41746.service - OpenSSH per-connection server daemon (10.0.0.1:41746). Jan 21 05:56:39.123195 systemd-networkd[1528]: vxlan.calico: Gained IPv6LL Jan 21 05:56:39.130946 systemd-networkd[1528]: cali106c2b932fa: Gained IPv6LL Jan 21 05:56:39.160277 containerd[1639]: time="2026-01-21T05:56:39.159993253Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:39.167210 containerd[1639]: time="2026-01-21T05:56:39.163075397Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 21 05:56:39.167210 containerd[1639]: time="2026-01-21T05:56:39.163598184Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:39.168805 kubelet[2982]: E0121 05:56:39.168026 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:39.168805 kubelet[2982]: E0121 05:56:39.168594 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:39.177945 kubelet[2982]: E0121 05:56:39.174808 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wclgb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:39.181754 kubelet[2982]: E0121 05:56:39.178719 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:56:39.290505 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2153387402.mount: Deactivated successfully. Jan 21 05:56:39.372947 containerd[1639]: time="2026-01-21T05:56:39.366788423Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zftdp,Uid:487c6aa2-04b5-4684-9420-0e4b73799fd3,Namespace:calico-system,Attempt:0,} returns sandbox id \"28a0e9bfa1bb48e5a590f98fba84659e25a1666df8cb2960d5ee3b2a2a8dee2c\"" Jan 21 05:56:39.423182 containerd[1639]: time="2026-01-21T05:56:39.422838542Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 21 05:56:39.430000 audit: BPF prog-id=213 op=LOAD Jan 21 05:56:39.430000 audit[6128]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff43e3a880 a2=98 a3=0 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.430000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.430000 audit: BPF prog-id=213 op=UNLOAD Jan 21 05:56:39.430000 audit[6128]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7fff43e3a850 a3=0 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.430000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.430000 audit: BPF prog-id=214 op=LOAD Jan 21 05:56:39.430000 audit[6128]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff43e3a690 a2=94 a3=54428f items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.430000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.430000 audit: BPF prog-id=214 op=UNLOAD Jan 21 05:56:39.430000 audit[6128]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7fff43e3a690 a2=94 a3=54428f items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.430000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.435998 kubelet[2982]: E0121 05:56:39.421109 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:56:39.436228 containerd[1639]: time="2026-01-21T05:56:39.428445801Z" level=info msg="Container 4c598845cbb7a2e325bd89688a7dea64e9388629fafbe181b34e230d833ac895: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:56:39.438078 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2291101085.mount: Deactivated successfully. Jan 21 05:56:39.430000 audit: BPF prog-id=215 op=LOAD Jan 21 05:56:39.430000 audit[6128]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff43e3a6c0 a2=94 a3=2 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.430000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.444000 audit: BPF prog-id=215 op=UNLOAD Jan 21 05:56:39.444000 audit[6128]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7fff43e3a6c0 a2=0 a3=2 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.444000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.444000 audit: BPF prog-id=216 op=LOAD Jan 21 05:56:39.444000 audit[6128]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff43e3a470 a2=94 a3=4 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.444000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.444000 audit: BPF prog-id=216 op=UNLOAD Jan 21 05:56:39.444000 audit[6128]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fff43e3a470 a2=94 a3=4 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.444000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.444000 audit: BPF prog-id=217 op=LOAD Jan 21 05:56:39.444000 audit[6128]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff43e3a570 a2=94 a3=7fff43e3a6f0 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.444000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.444000 audit: BPF prog-id=217 op=UNLOAD Jan 21 05:56:39.444000 audit[6128]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fff43e3a570 a2=0 a3=7fff43e3a6f0 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.444000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.479000 audit: BPF prog-id=218 op=LOAD Jan 21 05:56:39.479000 audit[6128]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff43e39ca0 a2=94 a3=2 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.479000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.482000 audit: BPF prog-id=218 op=UNLOAD Jan 21 05:56:39.482000 audit[6128]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fff43e39ca0 a2=0 a3=2 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.482000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.482000 audit: BPF prog-id=219 op=LOAD Jan 21 05:56:39.482000 audit: BPF prog-id=220 op=LOAD Jan 21 05:56:39.482000 audit[6128]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff43e39da0 a2=94 a3=30 items=0 ppid=5422 pid=6128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.482000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 21 05:56:39.483000 audit: BPF prog-id=221 op=LOAD Jan 21 05:56:39.483000 audit[6080]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00016e238 a2=98 a3=0 items=0 ppid=6054 pid=6080 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.483000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532376539626632313732303833666532383666346462316131343563 Jan 21 05:56:39.483000 audit: BPF prog-id=221 op=UNLOAD Jan 21 05:56:39.483000 audit[6080]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=6054 pid=6080 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.483000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532376539626632313732303833666532383666346462316131343563 Jan 21 05:56:39.489000 audit: BPF prog-id=222 op=LOAD Jan 21 05:56:39.489000 audit[6080]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00016e488 a2=98 a3=0 items=0 ppid=6054 pid=6080 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.489000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532376539626632313732303833666532383666346462316131343563 Jan 21 05:56:39.513000 audit: BPF prog-id=223 op=LOAD Jan 21 05:56:39.513000 audit[6080]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00016e218 a2=98 a3=0 items=0 ppid=6054 pid=6080 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.513000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532376539626632313732303833666532383666346462316131343563 Jan 21 05:56:39.513000 audit: BPF prog-id=223 op=UNLOAD Jan 21 05:56:39.513000 audit[6080]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=6054 pid=6080 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.513000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532376539626632313732303833666532383666346462316131343563 Jan 21 05:56:39.513000 audit: BPF prog-id=222 op=UNLOAD Jan 21 05:56:39.513000 audit[6080]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=6054 pid=6080 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.513000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532376539626632313732303833666532383666346462316131343563 Jan 21 05:56:39.528456 containerd[1639]: time="2026-01-21T05:56:39.528196811Z" level=info msg="CreateContainer within sandbox \"9aacbc556d4449617308332beda7a3d20077a51f0a64fa14368066d9e265a0f7\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"4c598845cbb7a2e325bd89688a7dea64e9388629fafbe181b34e230d833ac895\"" Jan 21 05:56:39.529452 containerd[1639]: time="2026-01-21T05:56:39.529264146Z" level=info msg="StartContainer for \"4c598845cbb7a2e325bd89688a7dea64e9388629fafbe181b34e230d833ac895\"" Jan 21 05:56:39.513000 audit: BPF prog-id=224 op=LOAD Jan 21 05:56:39.513000 audit[6080]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00016e6e8 a2=98 a3=0 items=0 ppid=6054 pid=6080 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.513000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532376539626632313732303833666532383666346462316131343563 Jan 21 05:56:39.546882 containerd[1639]: time="2026-01-21T05:56:39.546258249Z" level=info msg="connecting to shim 4c598845cbb7a2e325bd89688a7dea64e9388629fafbe181b34e230d833ac895" address="unix:///run/containerd/s/f8187a8e8e6b66fda912785ed062d55da4a4b3547deb157ace276fddd200645a" protocol=ttrpc version=3 Jan 21 05:56:39.547455 systemd-networkd[1528]: cali05f854b7b91: Link UP Jan 21 05:56:39.551076 systemd-networkd[1528]: cali05f854b7b91: Gained carrier Jan 21 05:56:39.617918 systemd-resolved[1290]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 21 05:56:39.637464 containerd[1639]: time="2026-01-21T05:56:39.635940639Z" level=info msg="connecting to shim 01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661" address="unix:///run/containerd/s/ea9dce5822c2b7f941d0e72d5fb88c97ff628e1d759e14a66de7924c6098e50f" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:56:39.703000 audit: BPF prog-id=225 op=LOAD Jan 21 05:56:39.703000 audit[6158]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe61e40ca0 a2=98 a3=0 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.703000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:39.706000 audit: BPF prog-id=225 op=UNLOAD Jan 21 05:56:39.706000 audit[6158]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffe61e40c70 a3=0 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.706000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:39.706000 audit: BPF prog-id=226 op=LOAD Jan 21 05:56:39.706000 audit[6158]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe61e40a90 a2=94 a3=54428f items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.706000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:39.707000 audit: BPF prog-id=226 op=UNLOAD Jan 21 05:56:39.707000 audit[6158]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe61e40a90 a2=94 a3=54428f items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.707000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:39.707000 audit: BPF prog-id=227 op=LOAD Jan 21 05:56:39.707000 audit[6158]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe61e40ac0 a2=94 a3=2 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.707000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:39.707000 audit: BPF prog-id=227 op=UNLOAD Jan 21 05:56:39.707000 audit[6158]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe61e40ac0 a2=0 a3=2 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.707000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:37.480 [INFO][5846] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--796779bb--6bz64-eth0 calico-apiserver-796779bb- calico-apiserver 137d495d-bc8e-4cd7-b462-b18325307853 1105 0 2026-01-21 05:54:00 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:796779bb projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-796779bb-6bz64 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali05f854b7b91 [] [] }} ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-6bz64" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--6bz64-" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:37.480 [INFO][5846] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-6bz64" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.541 [INFO][5957] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" HandleID="k8s-pod-network.05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Workload="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.566 [INFO][5957] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" HandleID="k8s-pod-network.05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Workload="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001960b0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-796779bb-6bz64", "timestamp":"2026-01-21 05:56:38.541603361 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.566 [INFO][5957] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.566 [INFO][5957] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.566 [INFO][5957] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.623 [INFO][5957] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" host="localhost" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.745 [INFO][5957] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.784 [INFO][5957] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.860 [INFO][5957] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.938 [INFO][5957] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.938 [INFO][5957] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" host="localhost" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:38.980 [INFO][5957] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1 Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:39.052 [INFO][5957] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" host="localhost" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:39.151 [INFO][5957] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" host="localhost" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:39.151 [INFO][5957] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" host="localhost" Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:39.151 [INFO][5957] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 21 05:56:39.712919 containerd[1639]: 2026-01-21 05:56:39.151 [INFO][5957] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" HandleID="k8s-pod-network.05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Workload="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" Jan 21 05:56:39.731486 containerd[1639]: 2026-01-21 05:56:39.207 [INFO][5846] cni-plugin/k8s.go 418: Populated endpoint ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-6bz64" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--796779bb--6bz64-eth0", GenerateName:"calico-apiserver-796779bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"137d495d-bc8e-4cd7-b462-b18325307853", ResourceVersion:"1105", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"796779bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-796779bb-6bz64", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali05f854b7b91", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:39.731486 containerd[1639]: 2026-01-21 05:56:39.207 [INFO][5846] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-6bz64" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" Jan 21 05:56:39.731486 containerd[1639]: 2026-01-21 05:56:39.207 [INFO][5846] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali05f854b7b91 ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-6bz64" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" Jan 21 05:56:39.731486 containerd[1639]: 2026-01-21 05:56:39.558 [INFO][5846] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-6bz64" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" Jan 21 05:56:39.731486 containerd[1639]: 2026-01-21 05:56:39.559 [INFO][5846] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-6bz64" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--796779bb--6bz64-eth0", GenerateName:"calico-apiserver-796779bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"137d495d-bc8e-4cd7-b462-b18325307853", ResourceVersion:"1105", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"796779bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1", Pod:"calico-apiserver-796779bb-6bz64", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali05f854b7b91", MAC:"42:9f:7e:58:5e:b7", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:39.731486 containerd[1639]: 2026-01-21 05:56:39.627 [INFO][5846] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" Namespace="calico-apiserver" Pod="calico-apiserver-796779bb-6bz64" WorkloadEndpoint="localhost-k8s-calico--apiserver--796779bb--6bz64-eth0" Jan 21 05:56:39.748000 audit[6159]: NETFILTER_CFG table=filter:125 family=2 entries=20 op=nft_register_rule pid=6159 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:39.748000 audit[6159]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc5e5c6ce0 a2=0 a3=7ffc5e5c6ccc items=0 ppid=3144 pid=6159 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.748000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:39.763000 audit[6159]: NETFILTER_CFG table=nat:126 family=2 entries=14 op=nft_register_rule pid=6159 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:39.763000 audit[6159]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffc5e5c6ce0 a2=0 a3=0 items=0 ppid=3144 pid=6159 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:39.763000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:39.833263 containerd[1639]: time="2026-01-21T05:56:39.832928062Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:39.860621 containerd[1639]: time="2026-01-21T05:56:39.859959441Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 21 05:56:39.860621 containerd[1639]: time="2026-01-21T05:56:39.860402630Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:39.864772 kubelet[2982]: E0121 05:56:39.863617 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 21 05:56:39.864772 kubelet[2982]: E0121 05:56:39.863878 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 21 05:56:39.870527 kubelet[2982]: E0121 05:56:39.870271 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgtqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:39.892038 containerd[1639]: time="2026-01-21T05:56:39.891124539Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 21 05:56:39.892856 systemd-networkd[1528]: cali402e0971676: Gained IPv6LL Jan 21 05:56:40.012264 containerd[1639]: time="2026-01-21T05:56:40.011965803Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:40.029030 containerd[1639]: time="2026-01-21T05:56:40.028514979Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 21 05:56:40.031000 audit[6105]: USER_ACCT pid=6105 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:40.032545 sshd[6105]: Accepted publickey for core from 10.0.0.1 port 41746 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:40.033552 kubelet[2982]: E0121 05:56:40.031180 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 21 05:56:40.033552 kubelet[2982]: E0121 05:56:40.031245 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 21 05:56:40.033552 kubelet[2982]: E0121 05:56:40.031474 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgtqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:40.033552 kubelet[2982]: E0121 05:56:40.033048 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:56:40.046000 audit[6105]: CRED_ACQ pid=6105 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:40.046000 audit[6105]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcab46fa60 a2=3 a3=0 items=0 ppid=1 pid=6105 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.046000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:40.051037 containerd[1639]: time="2026-01-21T05:56:40.031257106Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:40.055089 sshd-session[6105]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:40.092811 systemd-logind[1620]: New session 20 of user core. Jan 21 05:56:40.107936 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 21 05:56:40.168000 audit[6105]: USER_START pid=6105 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:40.184000 audit[6207]: CRED_ACQ pid=6207 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:40.191971 containerd[1639]: time="2026-01-21T05:56:40.191924982Z" level=info msg="connecting to shim 05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1" address="unix:///run/containerd/s/02e318acb396e58234149d1d496cc7c722cc08d59e72c1eeed501a08075ed11f" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:56:40.249549 systemd[1]: Started cri-containerd-4c598845cbb7a2e325bd89688a7dea64e9388629fafbe181b34e230d833ac895.scope - libcontainer container 4c598845cbb7a2e325bd89688a7dea64e9388629fafbe181b34e230d833ac895. Jan 21 05:56:40.292524 systemd[1]: Started cri-containerd-01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661.scope - libcontainer container 01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661. Jan 21 05:56:40.341562 containerd[1639]: time="2026-01-21T05:56:40.341377460Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-m7pjj,Uid:59842dba-2f25-4ec6-86e9-919aa2afb77a,Namespace:calico-system,Attempt:0,} returns sandbox id \"e27e9bf2172083fe286f4db1a145cbd79006660218bd088361fe68c4d2e38e6c\"" Jan 21 05:56:40.394819 containerd[1639]: time="2026-01-21T05:56:40.391503390Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 21 05:56:40.513570 kubelet[2982]: E0121 05:56:40.509461 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:56:40.544769 kubelet[2982]: E0121 05:56:40.529631 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:56:40.567000 audit: BPF prog-id=228 op=LOAD Jan 21 05:56:40.574000 audit: BPF prog-id=229 op=LOAD Jan 21 05:56:40.574000 audit[6167]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=6143 pid=6167 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.574000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3031373130303632643838346461666230353863653535306132643862 Jan 21 05:56:40.574000 audit: BPF prog-id=229 op=UNLOAD Jan 21 05:56:40.574000 audit[6167]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=6143 pid=6167 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.574000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3031373130303632643838346461666230353863653535306132643862 Jan 21 05:56:40.584000 audit: BPF prog-id=230 op=LOAD Jan 21 05:56:40.584000 audit[6167]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=6143 pid=6167 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.584000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3031373130303632643838346461666230353863653535306132643862 Jan 21 05:56:40.584000 audit: BPF prog-id=231 op=LOAD Jan 21 05:56:40.584000 audit[6167]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=6143 pid=6167 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.584000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3031373130303632643838346461666230353863653535306132643862 Jan 21 05:56:40.584000 audit: BPF prog-id=231 op=UNLOAD Jan 21 05:56:40.584000 audit[6167]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=6143 pid=6167 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.584000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3031373130303632643838346461666230353863653535306132643862 Jan 21 05:56:40.584000 audit: BPF prog-id=230 op=UNLOAD Jan 21 05:56:40.584000 audit[6167]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=6143 pid=6167 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.584000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3031373130303632643838346461666230353863653535306132643862 Jan 21 05:56:40.592000 audit: BPF prog-id=232 op=LOAD Jan 21 05:56:40.592000 audit[6167]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=6143 pid=6167 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.592000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3031373130303632643838346461666230353863653535306132643862 Jan 21 05:56:40.649000 audit: BPF prog-id=233 op=LOAD Jan 21 05:56:40.650922 systemd-resolved[1290]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 21 05:56:40.662000 audit: BPF prog-id=234 op=LOAD Jan 21 05:56:40.662000 audit[6152]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000160238 a2=98 a3=0 items=0 ppid=5907 pid=6152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.662000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3463353938383435636262376132653332356264383936383861376465 Jan 21 05:56:40.692000 audit: BPF prog-id=234 op=UNLOAD Jan 21 05:56:40.692000 audit[6152]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5907 pid=6152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.692000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3463353938383435636262376132653332356264383936383861376465 Jan 21 05:56:40.702000 audit: BPF prog-id=235 op=LOAD Jan 21 05:56:40.702000 audit[6152]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000160488 a2=98 a3=0 items=0 ppid=5907 pid=6152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.702000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3463353938383435636262376132653332356264383936383861376465 Jan 21 05:56:40.702000 audit: BPF prog-id=236 op=LOAD Jan 21 05:56:40.702000 audit[6152]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000160218 a2=98 a3=0 items=0 ppid=5907 pid=6152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.702000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3463353938383435636262376132653332356264383936383861376465 Jan 21 05:56:40.702000 audit: BPF prog-id=236 op=UNLOAD Jan 21 05:56:40.702000 audit[6152]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5907 pid=6152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.702000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3463353938383435636262376132653332356264383936383861376465 Jan 21 05:56:40.702000 audit: BPF prog-id=235 op=UNLOAD Jan 21 05:56:40.702000 audit[6152]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5907 pid=6152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.702000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3463353938383435636262376132653332356264383936383861376465 Jan 21 05:56:40.702000 audit: BPF prog-id=237 op=LOAD Jan 21 05:56:40.702000 audit[6152]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001606e8 a2=98 a3=0 items=0 ppid=5907 pid=6152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:40.702000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3463353938383435636262376132653332356264383936383861376465 Jan 21 05:56:40.730410 containerd[1639]: time="2026-01-21T05:56:40.730278254Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:40.748074 containerd[1639]: time="2026-01-21T05:56:40.747999726Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 21 05:56:40.749543 containerd[1639]: time="2026-01-21T05:56:40.748266504Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:40.754869 systemd-networkd[1528]: cali05f854b7b91: Gained IPv6LL Jan 21 05:56:40.761213 kubelet[2982]: E0121 05:56:40.757951 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 21 05:56:40.761213 kubelet[2982]: E0121 05:56:40.758022 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 21 05:56:40.761213 kubelet[2982]: E0121 05:56:40.758383 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6x2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:40.772421 kubelet[2982]: E0121 05:56:40.762222 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:56:40.781757 systemd-networkd[1528]: cali9e49729ca57: Link UP Jan 21 05:56:40.785390 systemd-networkd[1528]: cali9e49729ca57: Gained carrier Jan 21 05:56:40.882791 sshd[6207]: Connection closed by 10.0.0.1 port 41746 Jan 21 05:56:40.891000 audit[6105]: USER_END pid=6105 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:40.891000 audit[6105]: CRED_DISP pid=6105 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:40.881894 sshd-session[6105]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:40.909507 systemd[1]: Started cri-containerd-05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1.scope - libcontainer container 05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1. Jan 21 05:56:40.912068 systemd[1]: sshd@18-10.0.0.61:22-10.0.0.1:41746.service: Deactivated successfully. Jan 21 05:56:40.912000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.61:22-10.0.0.1:41746 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:40.927863 systemd[1]: session-20.scope: Deactivated successfully. Jan 21 05:56:41.000269 systemd-logind[1620]: Session 20 logged out. Waiting for processes to exit. Jan 21 05:56:41.021537 systemd-logind[1620]: Removed session 20. Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:38.606 [INFO][6032] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0 calico-kube-controllers-7598d7cc9- calico-system 821a7ed1-bcf9-475e-908c-afcf62571645 1109 0 2026-01-21 05:54:34 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:7598d7cc9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-7598d7cc9-qv4mq eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali9e49729ca57 [] [] }} ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Namespace="calico-system" Pod="calico-kube-controllers-7598d7cc9-qv4mq" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:38.606 [INFO][6032] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Namespace="calico-system" Pod="calico-kube-controllers-7598d7cc9-qv4mq" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.028 [INFO][6086] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" HandleID="k8s-pod-network.947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Workload="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.028 [INFO][6086] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" HandleID="k8s-pod-network.947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Workload="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004e490), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-7598d7cc9-qv4mq", "timestamp":"2026-01-21 05:56:39.028262573 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.030 [INFO][6086] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.188 [INFO][6086] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.189 [INFO][6086] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.392 [INFO][6086] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" host="localhost" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.846 [INFO][6086] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.935 [INFO][6086] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.958 [INFO][6086] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.993 [INFO][6086] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:39.998 [INFO][6086] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" host="localhost" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:40.027 [INFO][6086] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800 Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:40.137 [INFO][6086] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" host="localhost" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:40.315 [INFO][6086] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.137/26] block=192.168.88.128/26 handle="k8s-pod-network.947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" host="localhost" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:40.321 [INFO][6086] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.137/26] handle="k8s-pod-network.947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" host="localhost" Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:40.340 [INFO][6086] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 21 05:56:41.023193 containerd[1639]: 2026-01-21 05:56:40.340 [INFO][6086] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.137/26] IPv6=[] ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" HandleID="k8s-pod-network.947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Workload="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" Jan 21 05:56:41.032905 containerd[1639]: 2026-01-21 05:56:40.420 [INFO][6032] cni-plugin/k8s.go 418: Populated endpoint ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Namespace="calico-system" Pod="calico-kube-controllers-7598d7cc9-qv4mq" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0", GenerateName:"calico-kube-controllers-7598d7cc9-", Namespace:"calico-system", SelfLink:"", UID:"821a7ed1-bcf9-475e-908c-afcf62571645", ResourceVersion:"1109", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7598d7cc9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-7598d7cc9-qv4mq", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.137/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali9e49729ca57", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:41.032905 containerd[1639]: 2026-01-21 05:56:40.422 [INFO][6032] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.137/32] ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Namespace="calico-system" Pod="calico-kube-controllers-7598d7cc9-qv4mq" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" Jan 21 05:56:41.032905 containerd[1639]: 2026-01-21 05:56:40.422 [INFO][6032] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9e49729ca57 ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Namespace="calico-system" Pod="calico-kube-controllers-7598d7cc9-qv4mq" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" Jan 21 05:56:41.032905 containerd[1639]: 2026-01-21 05:56:40.789 [INFO][6032] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Namespace="calico-system" Pod="calico-kube-controllers-7598d7cc9-qv4mq" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" Jan 21 05:56:41.032905 containerd[1639]: 2026-01-21 05:56:40.797 [INFO][6032] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Namespace="calico-system" Pod="calico-kube-controllers-7598d7cc9-qv4mq" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0", GenerateName:"calico-kube-controllers-7598d7cc9-", Namespace:"calico-system", SelfLink:"", UID:"821a7ed1-bcf9-475e-908c-afcf62571645", ResourceVersion:"1109", Generation:0, CreationTimestamp:time.Date(2026, time.January, 21, 5, 54, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7598d7cc9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800", Pod:"calico-kube-controllers-7598d7cc9-qv4mq", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.137/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali9e49729ca57", MAC:"2a:59:88:4f:81:de", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 21 05:56:41.032905 containerd[1639]: 2026-01-21 05:56:40.855 [INFO][6032] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" Namespace="calico-system" Pod="calico-kube-controllers-7598d7cc9-qv4mq" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7598d7cc9--qv4mq-eth0" Jan 21 05:56:41.125000 audit: BPF prog-id=238 op=LOAD Jan 21 05:56:41.146163 kernel: kauditd_printk_skb: 256 callbacks suppressed Jan 21 05:56:41.146396 kernel: audit: type=1334 audit(1768975001.125:789): prog-id=238 op=LOAD Jan 21 05:56:41.129000 audit: BPF prog-id=239 op=LOAD Jan 21 05:56:41.129000 audit[6241]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00016e238 a2=98 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.182202 systemd-resolved[1290]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 21 05:56:41.271441 kernel: audit: type=1334 audit(1768975001.129:790): prog-id=239 op=LOAD Jan 21 05:56:41.271591 kernel: audit: type=1300 audit(1768975001.129:790): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00016e238 a2=98 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.271792 kernel: audit: type=1327 audit(1768975001.129:790): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.129000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.345210 kernel: audit: type=1334 audit(1768975001.129:791): prog-id=239 op=UNLOAD Jan 21 05:56:41.129000 audit: BPF prog-id=239 op=UNLOAD Jan 21 05:56:41.393414 kernel: audit: type=1300 audit(1768975001.129:791): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.129000 audit[6241]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.129000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.432177 kernel: audit: type=1327 audit(1768975001.129:791): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.451527 containerd[1639]: time="2026-01-21T05:56:41.448628664Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-qpdqz,Uid:9bc2381f-9591-401f-8403-a07aa98276a8,Namespace:kube-system,Attempt:0,} returns sandbox id \"01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661\"" Jan 21 05:56:41.129000 audit: BPF prog-id=240 op=LOAD Jan 21 05:56:41.461731 kernel: audit: type=1334 audit(1768975001.129:792): prog-id=240 op=LOAD Jan 21 05:56:41.528803 kernel: audit: type=1300 audit(1768975001.129:792): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00016e488 a2=98 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.129000 audit[6241]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00016e488 a2=98 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.529268 containerd[1639]: time="2026-01-21T05:56:41.518169674Z" level=info msg="CreateContainer within sandbox \"01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 21 05:56:41.529437 kubelet[2982]: E0121 05:56:41.478996 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:41.129000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.129000 audit: BPF prog-id=241 op=LOAD Jan 21 05:56:41.129000 audit[6241]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00016e218 a2=98 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.129000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.601787 kernel: audit: type=1327 audit(1768975001.129:792): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.129000 audit: BPF prog-id=241 op=UNLOAD Jan 21 05:56:41.129000 audit[6241]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.129000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.129000 audit: BPF prog-id=240 op=UNLOAD Jan 21 05:56:41.129000 audit[6241]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.129000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.129000 audit: BPF prog-id=242 op=LOAD Jan 21 05:56:41.129000 audit[6241]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00016e6e8 a2=98 a3=0 items=0 ppid=6208 pid=6241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.129000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035656664663638616332373361326436633039326538356431353463 Jan 21 05:56:41.173000 audit: BPF prog-id=243 op=LOAD Jan 21 05:56:41.173000 audit[6158]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe61e40980 a2=94 a3=1 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.173000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.174000 audit: BPF prog-id=243 op=UNLOAD Jan 21 05:56:41.174000 audit[6158]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe61e40980 a2=94 a3=1 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.174000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.195000 audit: BPF prog-id=244 op=LOAD Jan 21 05:56:41.195000 audit[6158]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe61e40970 a2=94 a3=4 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.195000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.195000 audit: BPF prog-id=244 op=UNLOAD Jan 21 05:56:41.195000 audit[6158]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe61e40970 a2=0 a3=4 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.195000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.195000 audit: BPF prog-id=245 op=LOAD Jan 21 05:56:41.195000 audit[6158]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffe61e407d0 a2=94 a3=5 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.195000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.195000 audit: BPF prog-id=245 op=UNLOAD Jan 21 05:56:41.195000 audit[6158]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffe61e407d0 a2=0 a3=5 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.195000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.195000 audit: BPF prog-id=246 op=LOAD Jan 21 05:56:41.195000 audit[6158]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe61e409f0 a2=94 a3=6 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.195000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.195000 audit: BPF prog-id=246 op=UNLOAD Jan 21 05:56:41.195000 audit[6158]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe61e409f0 a2=0 a3=6 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.195000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.195000 audit: BPF prog-id=247 op=LOAD Jan 21 05:56:41.195000 audit[6158]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe61e401a0 a2=94 a3=88 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.195000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.195000 audit: BPF prog-id=248 op=LOAD Jan 21 05:56:41.195000 audit[6158]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7ffe61e40020 a2=94 a3=2 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.195000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.603000 audit: BPF prog-id=248 op=UNLOAD Jan 21 05:56:41.603000 audit[6158]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7ffe61e40050 a2=0 a3=7ffe61e40150 items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.603000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.604000 audit: BPF prog-id=247 op=UNLOAD Jan 21 05:56:41.604000 audit[6158]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=341e3d10 a2=0 a3=71eababc80aea84c items=0 ppid=5422 pid=6158 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.604000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 21 05:56:41.633828 kubelet[2982]: E0121 05:56:41.628055 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:56:41.752773 containerd[1639]: time="2026-01-21T05:56:41.751417814Z" level=info msg="connecting to shim 947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800" address="unix:///run/containerd/s/b564dd6857527f2c47fbfab6fd558ba0abc2be0d250474d6b2d2ca7f1706fa04" namespace=k8s.io protocol=ttrpc version=3 Jan 21 05:56:41.804413 containerd[1639]: time="2026-01-21T05:56:41.785943535Z" level=info msg="StartContainer for \"4c598845cbb7a2e325bd89688a7dea64e9388629fafbe181b34e230d833ac895\" returns successfully" Jan 21 05:56:41.856848 kubelet[2982]: E0121 05:56:41.834204 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:56:41.866618 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3649839896.mount: Deactivated successfully. Jan 21 05:56:41.911000 audit[6303]: NETFILTER_CFG table=filter:127 family=2 entries=20 op=nft_register_rule pid=6303 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:41.911000 audit[6303]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7fff92874370 a2=0 a3=7fff9287435c items=0 ppid=3144 pid=6303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:41.911000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:42.011028 containerd[1639]: time="2026-01-21T05:56:42.010274125Z" level=info msg="Container ddd770e2ed108bb585707ff512e56dc40c3e8f539abdf252c51274b3d35d75c5: CDI devices from CRI Config.CDIDevices: []" Jan 21 05:56:42.048208 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1658655198.mount: Deactivated successfully. Jan 21 05:56:42.089000 audit[6303]: NETFILTER_CFG table=nat:128 family=2 entries=14 op=nft_register_rule pid=6303 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:42.089000 audit[6303]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7fff92874370 a2=0 a3=0 items=0 ppid=3144 pid=6303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.089000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:42.245000 audit: BPF prog-id=220 op=UNLOAD Jan 21 05:56:42.245000 audit[5422]: SYSCALL arch=c000003e syscall=263 success=yes exit=0 a0=ffffffffffffff9c a1=c001157240 a2=0 a3=0 items=0 ppid=5414 pid=5422 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="calico-node" exe="/usr/bin/calico-node" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.245000 audit: PROCTITLE proctitle=63616C69636F2D6E6F6465002D66656C6978 Jan 21 05:56:42.277553 containerd[1639]: time="2026-01-21T05:56:42.277272872Z" level=info msg="CreateContainer within sandbox \"01710062d884dafb058ce550a2d8b50b578836c3727f26628b9465ce13ca8661\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ddd770e2ed108bb585707ff512e56dc40c3e8f539abdf252c51274b3d35d75c5\"" Jan 21 05:56:42.359631 containerd[1639]: time="2026-01-21T05:56:42.330015646Z" level=info msg="StartContainer for \"ddd770e2ed108bb585707ff512e56dc40c3e8f539abdf252c51274b3d35d75c5\"" Jan 21 05:56:42.359631 containerd[1639]: time="2026-01-21T05:56:42.351112190Z" level=info msg="connecting to shim ddd770e2ed108bb585707ff512e56dc40c3e8f539abdf252c51274b3d35d75c5" address="unix:///run/containerd/s/ea9dce5822c2b7f941d0e72d5fb88c97ff628e1d759e14a66de7924c6098e50f" protocol=ttrpc version=3 Jan 21 05:56:42.403548 systemd-networkd[1528]: cali9e49729ca57: Gained IPv6LL Jan 21 05:56:42.490188 containerd[1639]: time="2026-01-21T05:56:42.490058317Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-796779bb-6bz64,Uid:137d495d-bc8e-4cd7-b462-b18325307853,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"05efdf68ac273a2d6c092e85d154c5ef75548dc6e3d774103aba07348afbeed1\"" Jan 21 05:56:42.507466 containerd[1639]: time="2026-01-21T05:56:42.507288873Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 21 05:56:42.547204 systemd[1]: Started cri-containerd-ddd770e2ed108bb585707ff512e56dc40c3e8f539abdf252c51274b3d35d75c5.scope - libcontainer container ddd770e2ed108bb585707ff512e56dc40c3e8f539abdf252c51274b3d35d75c5. Jan 21 05:56:42.630569 systemd[1]: Started cri-containerd-947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800.scope - libcontainer container 947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800. Jan 21 05:56:42.668243 containerd[1639]: time="2026-01-21T05:56:42.668118052Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:42.679467 kubelet[2982]: E0121 05:56:42.679171 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:42.687543 containerd[1639]: time="2026-01-21T05:56:42.682831229Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 21 05:56:42.687543 containerd[1639]: time="2026-01-21T05:56:42.682950581Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:42.687828 kubelet[2982]: E0121 05:56:42.683599 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:42.687828 kubelet[2982]: E0121 05:56:42.686024 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:42.687828 kubelet[2982]: E0121 05:56:42.686251 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6nlrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:42.689286 kubelet[2982]: E0121 05:56:42.689252 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:56:42.755000 audit: BPF prog-id=249 op=LOAD Jan 21 05:56:42.756000 audit: BPF prog-id=250 op=LOAD Jan 21 05:56:42.756000 audit[6329]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c238 a2=98 a3=0 items=0 ppid=6143 pid=6329 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.756000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6464643737306532656431303862623538353730376666353132653536 Jan 21 05:56:42.756000 audit: BPF prog-id=250 op=UNLOAD Jan 21 05:56:42.756000 audit[6329]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=6143 pid=6329 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.756000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6464643737306532656431303862623538353730376666353132653536 Jan 21 05:56:42.775000 audit: BPF prog-id=251 op=LOAD Jan 21 05:56:42.775000 audit[6329]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c488 a2=98 a3=0 items=0 ppid=6143 pid=6329 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.775000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6464643737306532656431303862623538353730376666353132653536 Jan 21 05:56:42.778000 audit: BPF prog-id=252 op=LOAD Jan 21 05:56:42.778000 audit[6329]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00010c218 a2=98 a3=0 items=0 ppid=6143 pid=6329 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.778000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6464643737306532656431303862623538353730376666353132653536 Jan 21 05:56:42.778000 audit: BPF prog-id=252 op=UNLOAD Jan 21 05:56:42.778000 audit[6329]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=6143 pid=6329 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.778000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6464643737306532656431303862623538353730376666353132653536 Jan 21 05:56:42.778000 audit: BPF prog-id=251 op=UNLOAD Jan 21 05:56:42.778000 audit[6329]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=6143 pid=6329 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.778000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6464643737306532656431303862623538353730376666353132653536 Jan 21 05:56:42.778000 audit: BPF prog-id=253 op=LOAD Jan 21 05:56:42.778000 audit[6329]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00010c6e8 a2=98 a3=0 items=0 ppid=6143 pid=6329 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.778000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6464643737306532656431303862623538353730376666353132653536 Jan 21 05:56:42.927000 audit[6365]: NETFILTER_CFG table=filter:129 family=2 entries=20 op=nft_register_rule pid=6365 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:42.927000 audit[6365]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffe79f5f8f0 a2=0 a3=7ffe79f5f8dc items=0 ppid=3144 pid=6365 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.927000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:42.951000 audit[6365]: NETFILTER_CFG table=nat:130 family=2 entries=14 op=nft_register_rule pid=6365 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:42.951000 audit[6365]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffe79f5f8f0 a2=0 a3=0 items=0 ppid=3144 pid=6365 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:42.951000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:43.027316 containerd[1639]: time="2026-01-21T05:56:43.027004781Z" level=info msg="StartContainer for \"ddd770e2ed108bb585707ff512e56dc40c3e8f539abdf252c51274b3d35d75c5\" returns successfully" Jan 21 05:56:43.138000 audit: BPF prog-id=254 op=LOAD Jan 21 05:56:43.140000 audit: BPF prog-id=255 op=LOAD Jan 21 05:56:43.140000 audit[6318]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001da238 a2=98 a3=0 items=0 ppid=6296 pid=6318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:43.140000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3934376632646332656339666536356634343239613331373239326161 Jan 21 05:56:43.145000 audit: BPF prog-id=255 op=UNLOAD Jan 21 05:56:43.145000 audit[6318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=6296 pid=6318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:43.145000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3934376632646332656339666536356634343239613331373239326161 Jan 21 05:56:43.146000 audit: BPF prog-id=256 op=LOAD Jan 21 05:56:43.146000 audit[6318]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001da488 a2=98 a3=0 items=0 ppid=6296 pid=6318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:43.146000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3934376632646332656339666536356634343239613331373239326161 Jan 21 05:56:43.147000 audit: BPF prog-id=257 op=LOAD Jan 21 05:56:43.147000 audit[6318]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001da218 a2=98 a3=0 items=0 ppid=6296 pid=6318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:43.147000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3934376632646332656339666536356634343239613331373239326161 Jan 21 05:56:43.149000 audit: BPF prog-id=257 op=UNLOAD Jan 21 05:56:43.149000 audit[6318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=6296 pid=6318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:43.149000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3934376632646332656339666536356634343239613331373239326161 Jan 21 05:56:43.149000 audit: BPF prog-id=256 op=UNLOAD Jan 21 05:56:43.149000 audit[6318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=6296 pid=6318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:43.149000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3934376632646332656339666536356634343239613331373239326161 Jan 21 05:56:43.149000 audit: BPF prog-id=258 op=LOAD Jan 21 05:56:43.149000 audit[6318]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001da6e8 a2=98 a3=0 items=0 ppid=6296 pid=6318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:43.149000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3934376632646332656339666536356634343239613331373239326161 Jan 21 05:56:43.219964 systemd-resolved[1290]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 21 05:56:43.713955 kubelet[2982]: E0121 05:56:43.713914 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:43.729492 kubelet[2982]: E0121 05:56:43.720885 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:43.765021 kubelet[2982]: E0121 05:56:43.763432 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:56:43.946479 containerd[1639]: time="2026-01-21T05:56:43.940164200Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7598d7cc9-qv4mq,Uid:821a7ed1-bcf9-475e-908c-afcf62571645,Namespace:calico-system,Attempt:0,} returns sandbox id \"947f2dc2ec9fe65f4429a317292aa67201e546b927d2247cdc682727eb58b800\"" Jan 21 05:56:43.961775 kubelet[2982]: I0121 05:56:43.961257 2982 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-sgzh7" podStartSLOduration=194.961095071 podStartE2EDuration="3m14.961095071s" podCreationTimestamp="2026-01-21 05:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 05:56:42.755298068 +0000 UTC m=+196.964947801" watchObservedRunningTime="2026-01-21 05:56:43.961095071 +0000 UTC m=+198.170744774" Jan 21 05:56:43.966923 kubelet[2982]: I0121 05:56:43.964165 2982 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-qpdqz" podStartSLOduration=194.964149167 podStartE2EDuration="3m14.964149167s" podCreationTimestamp="2026-01-21 05:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 05:56:43.933818613 +0000 UTC m=+198.143468346" watchObservedRunningTime="2026-01-21 05:56:43.964149167 +0000 UTC m=+198.173798871" Jan 21 05:56:43.983271 containerd[1639]: time="2026-01-21T05:56:43.983210303Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 21 05:56:44.094915 containerd[1639]: time="2026-01-21T05:56:44.094252574Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:44.099970 containerd[1639]: time="2026-01-21T05:56:44.097747503Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 21 05:56:44.099970 containerd[1639]: time="2026-01-21T05:56:44.097861737Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:44.126804 kubelet[2982]: E0121 05:56:44.118867 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 21 05:56:44.126804 kubelet[2982]: E0121 05:56:44.118940 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 21 05:56:44.126804 kubelet[2982]: E0121 05:56:44.119194 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8n95t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:44.135413 kubelet[2982]: E0121 05:56:44.135292 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:56:44.261000 audit[6407]: NETFILTER_CFG table=filter:131 family=2 entries=20 op=nft_register_rule pid=6407 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:44.261000 audit[6407]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffcca6280c0 a2=0 a3=7ffcca6280ac items=0 ppid=3144 pid=6407 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:44.261000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:44.378000 audit[6407]: NETFILTER_CFG table=nat:132 family=2 entries=14 op=nft_register_rule pid=6407 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:44.378000 audit[6407]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffcca6280c0 a2=0 a3=0 items=0 ppid=3144 pid=6407 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:44.378000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:44.521000 audit[6414]: NETFILTER_CFG table=filter:133 family=2 entries=17 op=nft_register_rule pid=6414 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:44.521000 audit[6414]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fff7cde4cf0 a2=0 a3=7fff7cde4cdc items=0 ppid=3144 pid=6414 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:44.521000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:44.583000 audit[6414]: NETFILTER_CFG table=nat:134 family=2 entries=47 op=nft_register_chain pid=6414 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:56:44.583000 audit[6414]: SYSCALL arch=c000003e syscall=46 success=yes exit=19860 a0=3 a1=7fff7cde4cf0 a2=0 a3=7fff7cde4cdc items=0 ppid=3144 pid=6414 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:44.583000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:56:44.644000 audit[6423]: NETFILTER_CFG table=nat:135 family=2 entries=15 op=nft_register_chain pid=6423 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 21 05:56:44.644000 audit[6423]: SYSCALL arch=c000003e syscall=46 success=yes exit=5084 a0=3 a1=7ffcd2f80650 a2=0 a3=7ffcd2f8063c items=0 ppid=5422 pid=6423 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:44.644000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 21 05:56:44.672000 audit[6424]: NETFILTER_CFG table=mangle:136 family=2 entries=16 op=nft_register_chain pid=6424 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 21 05:56:44.672000 audit[6424]: SYSCALL arch=c000003e syscall=46 success=yes exit=6868 a0=3 a1=7ffcb9aa4af0 a2=0 a3=7ffcb9aa4adc items=0 ppid=5422 pid=6424 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:44.672000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 21 05:56:44.828000 audit[6426]: NETFILTER_CFG table=raw:137 family=2 entries=21 op=nft_register_chain pid=6426 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 21 05:56:44.828000 audit[6426]: SYSCALL arch=c000003e syscall=46 success=yes exit=8452 a0=3 a1=7fff2b9437b0 a2=0 a3=7fff2b94379c items=0 ppid=5422 pid=6426 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:44.828000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 21 05:56:44.861324 kubelet[2982]: E0121 05:56:44.861278 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:44.902840 kubelet[2982]: E0121 05:56:44.878791 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:44.926803 kubelet[2982]: E0121 05:56:44.922626 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:56:44.957000 audit[6428]: NETFILTER_CFG table=filter:138 family=2 entries=233 op=nft_register_chain pid=6428 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 21 05:56:44.957000 audit[6428]: SYSCALL arch=c000003e syscall=46 success=yes exit=136592 a0=3 a1=7ffc57dc63f0 a2=0 a3=7ffc57dc63dc items=0 ppid=5422 pid=6428 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:44.957000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 21 05:56:45.309000 audit[6438]: NETFILTER_CFG table=filter:139 family=2 entries=137 op=nft_register_chain pid=6438 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 21 05:56:45.309000 audit[6438]: SYSCALL arch=c000003e syscall=46 success=yes exit=79460 a0=3 a1=7ffd73172ad0 a2=0 a3=7ffd73172abc items=0 ppid=5422 pid=6438 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:45.309000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 21 05:56:45.890531 kubelet[2982]: E0121 05:56:45.888571 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:45.890531 kubelet[2982]: E0121 05:56:45.888937 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:56:45.987000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.61:22-10.0.0.1:57038 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:45.987493 systemd[1]: Started sshd@19-10.0.0.61:22-10.0.0.1:57038.service - OpenSSH per-connection server daemon (10.0.0.1:57038). Jan 21 05:56:46.334000 audit[6440]: USER_ACCT pid=6440 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:46.355081 kernel: kauditd_printk_skb: 135 callbacks suppressed Jan 21 05:56:46.355216 kernel: audit: type=1101 audit(1768975006.334:840): pid=6440 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:46.355285 sshd[6440]: Accepted publickey for core from 10.0.0.1 port 57038 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:46.368003 sshd-session[6440]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:46.357000 audit[6440]: CRED_ACQ pid=6440 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:46.429038 systemd-logind[1620]: New session 21 of user core. Jan 21 05:56:46.468612 kernel: audit: type=1103 audit(1768975006.357:841): pid=6440 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:46.468912 kernel: audit: type=1006 audit(1768975006.357:842): pid=6440 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=21 res=1 Jan 21 05:56:46.357000 audit[6440]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe113d4a10 a2=3 a3=0 items=0 ppid=1 pid=6440 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:46.516244 kernel: audit: type=1300 audit(1768975006.357:842): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe113d4a10 a2=3 a3=0 items=0 ppid=1 pid=6440 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:46.516503 kernel: audit: type=1327 audit(1768975006.357:842): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:46.357000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:46.521263 systemd[1]: Started session-21.scope - Session 21 of User core. Jan 21 05:56:46.552000 audit[6440]: USER_START pid=6440 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:46.607020 kernel: audit: type=1105 audit(1768975006.552:843): pid=6440 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:46.588000 audit[6444]: CRED_ACQ pid=6444 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:46.660794 kernel: audit: type=1103 audit(1768975006.588:844): pid=6444 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:47.192947 sshd[6444]: Connection closed by 10.0.0.1 port 57038 Jan 21 05:56:47.193985 sshd-session[6440]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:47.199000 audit[6440]: USER_END pid=6440 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:47.234274 containerd[1639]: time="2026-01-21T05:56:47.212930362Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 21 05:56:47.199000 audit[6440]: CRED_DISP pid=6440 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:47.263297 systemd[1]: sshd@19-10.0.0.61:22-10.0.0.1:57038.service: Deactivated successfully. Jan 21 05:56:47.275052 kernel: audit: type=1106 audit(1768975007.199:845): pid=6440 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:47.275146 kernel: audit: type=1104 audit(1768975007.199:846): pid=6440 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:47.275175 kernel: audit: type=1131 audit(1768975007.270:847): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.61:22-10.0.0.1:57038 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:47.270000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.61:22-10.0.0.1:57038 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:47.281808 systemd[1]: session-21.scope: Deactivated successfully. Jan 21 05:56:47.293855 systemd-logind[1620]: Session 21 logged out. Waiting for processes to exit. Jan 21 05:56:47.304307 systemd-logind[1620]: Removed session 21. Jan 21 05:56:47.367617 containerd[1639]: time="2026-01-21T05:56:47.367086840Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:47.373488 containerd[1639]: time="2026-01-21T05:56:47.372558419Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 21 05:56:47.373488 containerd[1639]: time="2026-01-21T05:56:47.372942026Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:47.383916 kubelet[2982]: E0121 05:56:47.382118 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:47.387111 kubelet[2982]: E0121 05:56:47.386619 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:47.392213 kubelet[2982]: E0121 05:56:47.389626 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n5wjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:47.397109 kubelet[2982]: E0121 05:56:47.397041 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:56:48.238596 kubelet[2982]: E0121 05:56:48.234879 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:56:50.223334 containerd[1639]: time="2026-01-21T05:56:50.222981948Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 21 05:56:50.323935 containerd[1639]: time="2026-01-21T05:56:50.323364375Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:50.342476 containerd[1639]: time="2026-01-21T05:56:50.341339151Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 21 05:56:50.342476 containerd[1639]: time="2026-01-21T05:56:50.341560434Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:50.348603 kubelet[2982]: E0121 05:56:50.347082 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 21 05:56:50.348603 kubelet[2982]: E0121 05:56:50.347155 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 21 05:56:50.348603 kubelet[2982]: E0121 05:56:50.347315 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:2d409fc375e643af9c78cec77dff7f0b,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rgkt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-cd95c566-k55c2_calico-system(da0d7215-59d8-4e4a-9787-c2942086bde3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:50.358505 containerd[1639]: time="2026-01-21T05:56:50.358318414Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 21 05:56:50.516140 containerd[1639]: time="2026-01-21T05:56:50.514285844Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:50.530222 containerd[1639]: time="2026-01-21T05:56:50.529943275Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 21 05:56:50.530222 containerd[1639]: time="2026-01-21T05:56:50.530145904Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:50.532291 kubelet[2982]: E0121 05:56:50.531100 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 21 05:56:50.532291 kubelet[2982]: E0121 05:56:50.531184 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 21 05:56:50.544044 kubelet[2982]: E0121 05:56:50.531354 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rgkt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-cd95c566-k55c2_calico-system(da0d7215-59d8-4e4a-9787-c2942086bde3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:50.544044 kubelet[2982]: E0121 05:56:50.537866 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-cd95c566-k55c2" podUID="da0d7215-59d8-4e4a-9787-c2942086bde3" Jan 21 05:56:52.247240 systemd[1]: Started sshd@20-10.0.0.61:22-10.0.0.1:57052.service - OpenSSH per-connection server daemon (10.0.0.1:57052). Jan 21 05:56:52.246000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.61:22-10.0.0.1:57052 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:52.306104 kernel: audit: type=1130 audit(1768975012.246:848): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.61:22-10.0.0.1:57052 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:52.568000 audit[6466]: USER_ACCT pid=6466 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:52.574879 sshd[6466]: Accepted publickey for core from 10.0.0.1 port 57052 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:52.581358 sshd-session[6466]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:52.616829 kernel: audit: type=1101 audit(1768975012.568:849): pid=6466 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:52.617020 kernel: audit: type=1103 audit(1768975012.578:850): pid=6466 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:52.578000 audit[6466]: CRED_ACQ pid=6466 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:52.635235 systemd-logind[1620]: New session 22 of user core. Jan 21 05:56:52.679878 kernel: audit: type=1006 audit(1768975012.578:851): pid=6466 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=22 res=1 Jan 21 05:56:52.680070 kernel: audit: type=1300 audit(1768975012.578:851): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc4895aa30 a2=3 a3=0 items=0 ppid=1 pid=6466 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:52.578000 audit[6466]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc4895aa30 a2=3 a3=0 items=0 ppid=1 pid=6466 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:52.740974 kernel: audit: type=1327 audit(1768975012.578:851): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:52.578000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:52.760019 systemd[1]: Started session-22.scope - Session 22 of User core. Jan 21 05:56:52.791000 audit[6466]: USER_START pid=6466 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:52.894842 kernel: audit: type=1105 audit(1768975012.791:852): pid=6466 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:52.895420 kernel: audit: type=1103 audit(1768975012.810:853): pid=6470 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:52.810000 audit[6470]: CRED_ACQ pid=6470 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:53.234209 containerd[1639]: time="2026-01-21T05:56:53.228584258Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 21 05:56:53.364524 containerd[1639]: time="2026-01-21T05:56:53.363941327Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:53.378797 containerd[1639]: time="2026-01-21T05:56:53.378145449Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 21 05:56:53.378797 containerd[1639]: time="2026-01-21T05:56:53.378270442Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:53.379168 kubelet[2982]: E0121 05:56:53.378918 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 21 05:56:53.379168 kubelet[2982]: E0121 05:56:53.378981 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 21 05:56:53.383053 kubelet[2982]: E0121 05:56:53.379155 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6x2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:53.383053 kubelet[2982]: E0121 05:56:53.382553 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:56:53.593122 sshd[6470]: Connection closed by 10.0.0.1 port 57052 Jan 21 05:56:53.599983 sshd-session[6466]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:53.604000 audit[6466]: USER_END pid=6466 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:53.643073 systemd[1]: sshd@20-10.0.0.61:22-10.0.0.1:57052.service: Deactivated successfully. Jan 21 05:56:53.665330 systemd[1]: session-22.scope: Deactivated successfully. Jan 21 05:56:53.674825 systemd-logind[1620]: Session 22 logged out. Waiting for processes to exit. Jan 21 05:56:53.677963 systemd-logind[1620]: Removed session 22. Jan 21 05:56:53.695176 kernel: audit: type=1106 audit(1768975013.604:854): pid=6466 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:53.695331 kernel: audit: type=1104 audit(1768975013.605:855): pid=6466 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:53.605000 audit[6466]: CRED_DISP pid=6466 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:53.644000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.61:22-10.0.0.1:57052 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:54.231905 containerd[1639]: time="2026-01-21T05:56:54.230930882Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 21 05:56:54.324066 containerd[1639]: time="2026-01-21T05:56:54.323860568Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:54.335183 containerd[1639]: time="2026-01-21T05:56:54.334994033Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 21 05:56:54.340915 containerd[1639]: time="2026-01-21T05:56:54.336071504Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:54.350055 kubelet[2982]: E0121 05:56:54.342104 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 21 05:56:54.350055 kubelet[2982]: E0121 05:56:54.342176 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 21 05:56:54.350380 kubelet[2982]: E0121 05:56:54.350327 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgtqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:54.361254 containerd[1639]: time="2026-01-21T05:56:54.360855802Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 21 05:56:54.461308 containerd[1639]: time="2026-01-21T05:56:54.461241599Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:54.472048 containerd[1639]: time="2026-01-21T05:56:54.469241185Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 21 05:56:54.472048 containerd[1639]: time="2026-01-21T05:56:54.470296151Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:54.473172 kubelet[2982]: E0121 05:56:54.473065 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:54.473923 kubelet[2982]: E0121 05:56:54.473134 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:54.478175 kubelet[2982]: E0121 05:56:54.478091 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wclgb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:54.480968 containerd[1639]: time="2026-01-21T05:56:54.480417026Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 21 05:56:54.485170 kubelet[2982]: E0121 05:56:54.480072 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:56:54.589424 containerd[1639]: time="2026-01-21T05:56:54.589359320Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:54.595422 containerd[1639]: time="2026-01-21T05:56:54.595271691Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 21 05:56:54.595422 containerd[1639]: time="2026-01-21T05:56:54.595378390Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:54.599933 kubelet[2982]: E0121 05:56:54.597938 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 21 05:56:54.599933 kubelet[2982]: E0121 05:56:54.599795 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 21 05:56:54.601834 kubelet[2982]: E0121 05:56:54.600317 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgtqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:54.604377 kubelet[2982]: E0121 05:56:54.603873 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:56:58.226280 containerd[1639]: time="2026-01-21T05:56:58.226229375Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 21 05:56:58.321445 containerd[1639]: time="2026-01-21T05:56:58.321233757Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:58.328145 containerd[1639]: time="2026-01-21T05:56:58.328104084Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:58.329798 containerd[1639]: time="2026-01-21T05:56:58.329144104Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 21 05:56:58.331429 kubelet[2982]: E0121 05:56:58.331281 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 21 05:56:58.332296 kubelet[2982]: E0121 05:56:58.331432 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 21 05:56:58.333265 kubelet[2982]: E0121 05:56:58.333164 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8n95t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:58.336356 kubelet[2982]: E0121 05:56:58.336259 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:56:58.344157 containerd[1639]: time="2026-01-21T05:56:58.342352550Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 21 05:56:58.446302 containerd[1639]: time="2026-01-21T05:56:58.443248184Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:56:58.450836 containerd[1639]: time="2026-01-21T05:56:58.449902887Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 21 05:56:58.457624 containerd[1639]: time="2026-01-21T05:56:58.450558319Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 21 05:56:58.466388 kubelet[2982]: E0121 05:56:58.463438 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:58.466388 kubelet[2982]: E0121 05:56:58.466003 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:56:58.467030 kubelet[2982]: E0121 05:56:58.466774 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6nlrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 21 05:56:58.472000 kubelet[2982]: E0121 05:56:58.467969 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:56:58.633767 systemd[1]: Started sshd@21-10.0.0.61:22-10.0.0.1:56410.service - OpenSSH per-connection server daemon (10.0.0.1:56410). Jan 21 05:56:58.648795 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:56:58.648916 kernel: audit: type=1130 audit(1768975018.633:857): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.61:22-10.0.0.1:56410 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:58.633000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.61:22-10.0.0.1:56410 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:56:58.945000 audit[6500]: USER_ACCT pid=6500 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:58.948259 sshd[6500]: Accepted publickey for core from 10.0.0.1 port 56410 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:56:58.961977 sshd-session[6500]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:56:58.995445 systemd-logind[1620]: New session 23 of user core. Jan 21 05:56:58.953000 audit[6500]: CRED_ACQ pid=6500 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.054588 kernel: audit: type=1101 audit(1768975018.945:858): pid=6500 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.054884 kernel: audit: type=1103 audit(1768975018.953:859): pid=6500 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.055932 kernel: audit: type=1006 audit(1768975018.953:860): pid=6500 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=23 res=1 Jan 21 05:56:58.953000 audit[6500]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc226f2bb0 a2=3 a3=0 items=0 ppid=1 pid=6500 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:59.094013 kernel: audit: type=1300 audit(1768975018.953:860): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc226f2bb0 a2=3 a3=0 items=0 ppid=1 pid=6500 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:56:58.953000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:59.145965 kernel: audit: type=1327 audit(1768975018.953:860): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:56:59.149370 systemd[1]: Started session-23.scope - Session 23 of User core. Jan 21 05:56:59.167000 audit[6500]: USER_START pid=6500 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.178000 audit[6504]: CRED_ACQ pid=6504 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.243051 kubelet[2982]: E0121 05:56:59.242001 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:56:59.253993 kernel: audit: type=1105 audit(1768975019.167:861): pid=6500 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.254140 kernel: audit: type=1103 audit(1768975019.178:862): pid=6504 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.656159 sshd[6504]: Connection closed by 10.0.0.1 port 56410 Jan 21 05:56:59.656235 sshd-session[6500]: pam_unix(sshd:session): session closed for user core Jan 21 05:56:59.661000 audit[6500]: USER_END pid=6500 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.671925 systemd-logind[1620]: Session 23 logged out. Waiting for processes to exit. Jan 21 05:56:59.677450 systemd[1]: sshd@21-10.0.0.61:22-10.0.0.1:56410.service: Deactivated successfully. Jan 21 05:56:59.696121 systemd[1]: session-23.scope: Deactivated successfully. Jan 21 05:56:59.708070 systemd-logind[1620]: Removed session 23. Jan 21 05:56:59.661000 audit[6500]: CRED_DISP pid=6500 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.748392 kernel: audit: type=1106 audit(1768975019.661:863): pid=6500 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.748875 kernel: audit: type=1104 audit(1768975019.661:864): pid=6500 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:56:59.680000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.61:22-10.0.0.1:56410 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:00.444353 kubelet[2982]: E0121 05:57:00.444136 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:57:03.210605 kubelet[2982]: E0121 05:57:03.209334 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:57:04.689000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.61:22-10.0.0.1:36484 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:04.691849 systemd[1]: Started sshd@22-10.0.0.61:22-10.0.0.1:36484.service - OpenSSH per-connection server daemon (10.0.0.1:36484). Jan 21 05:57:04.720038 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:57:04.720204 kernel: audit: type=1130 audit(1768975024.689:866): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.61:22-10.0.0.1:36484 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:04.851000 audit[6546]: USER_ACCT pid=6546 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:04.853963 sshd[6546]: Accepted publickey for core from 10.0.0.1 port 36484 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:04.858373 sshd-session[6546]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:04.854000 audit[6546]: CRED_ACQ pid=6546 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:04.895068 systemd-logind[1620]: New session 24 of user core. Jan 21 05:57:04.932209 kernel: audit: type=1101 audit(1768975024.851:867): pid=6546 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:04.932365 kernel: audit: type=1103 audit(1768975024.854:868): pid=6546 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:04.932418 kernel: audit: type=1006 audit(1768975024.854:869): pid=6546 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=24 res=1 Jan 21 05:57:04.951109 kernel: audit: type=1300 audit(1768975024.854:869): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffef5ac9910 a2=3 a3=0 items=0 ppid=1 pid=6546 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:04.854000 audit[6546]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffef5ac9910 a2=3 a3=0 items=0 ppid=1 pid=6546 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:04.854000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:05.006510 kernel: audit: type=1327 audit(1768975024.854:869): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:05.012764 systemd[1]: Started session-24.scope - Session 24 of User core. Jan 21 05:57:05.023000 audit[6546]: USER_START pid=6546 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:05.082123 kernel: audit: type=1105 audit(1768975025.023:870): pid=6546 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:05.082263 kernel: audit: type=1103 audit(1768975025.029:871): pid=6556 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:05.029000 audit[6556]: CRED_ACQ pid=6556 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:05.212272 kubelet[2982]: E0121 05:57:05.211756 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:57:05.217151 kubelet[2982]: E0121 05:57:05.216773 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-cd95c566-k55c2" podUID="da0d7215-59d8-4e4a-9787-c2942086bde3" Jan 21 05:57:05.433859 sshd[6556]: Connection closed by 10.0.0.1 port 36484 Jan 21 05:57:05.435866 sshd-session[6546]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:05.444000 audit[6546]: USER_END pid=6546 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:05.450372 systemd[1]: sshd@22-10.0.0.61:22-10.0.0.1:36484.service: Deactivated successfully. Jan 21 05:57:05.452116 systemd-logind[1620]: Session 24 logged out. Waiting for processes to exit. Jan 21 05:57:05.461102 systemd[1]: session-24.scope: Deactivated successfully. Jan 21 05:57:05.473361 systemd-logind[1620]: Removed session 24. Jan 21 05:57:05.496787 kernel: audit: type=1106 audit(1768975025.444:872): pid=6546 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:05.444000 audit[6546]: CRED_DISP pid=6546 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:05.450000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.61:22-10.0.0.1:36484 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:05.534205 kernel: audit: type=1104 audit(1768975025.444:873): pid=6546 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:07.227033 kubelet[2982]: E0121 05:57:07.223621 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:57:08.216118 kubelet[2982]: E0121 05:57:08.213519 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:57:09.210500 kubelet[2982]: E0121 05:57:09.210303 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:57:10.212114 kubelet[2982]: E0121 05:57:10.212058 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:57:10.217355 containerd[1639]: time="2026-01-21T05:57:10.215486889Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 21 05:57:10.330823 containerd[1639]: time="2026-01-21T05:57:10.330748678Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:57:10.343148 containerd[1639]: time="2026-01-21T05:57:10.343014339Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 21 05:57:10.346756 containerd[1639]: time="2026-01-21T05:57:10.343542592Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 21 05:57:10.350089 kubelet[2982]: E0121 05:57:10.349894 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:57:10.350089 kubelet[2982]: E0121 05:57:10.350043 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:57:10.350383 kubelet[2982]: E0121 05:57:10.350240 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n5wjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-796779bb-f2fds_calico-apiserver(dac6efad-1039-4242-b1c8-782ac61e5470): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 21 05:57:10.352102 kubelet[2982]: E0121 05:57:10.351932 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:57:10.484931 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:57:10.485068 kernel: audit: type=1130 audit(1768975030.471:875): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.61:22-10.0.0.1:36498 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:10.471000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.61:22-10.0.0.1:36498 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:10.472208 systemd[1]: Started sshd@23-10.0.0.61:22-10.0.0.1:36498.service - OpenSSH per-connection server daemon (10.0.0.1:36498). Jan 21 05:57:10.686000 audit[6579]: USER_ACCT pid=6579 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:10.695296 sshd-session[6579]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:10.719238 kernel: audit: type=1101 audit(1768975030.686:876): pid=6579 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:10.719359 sshd[6579]: Accepted publickey for core from 10.0.0.1 port 36498 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:10.689000 audit[6579]: CRED_ACQ pid=6579 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:10.744540 systemd-logind[1620]: New session 25 of user core. Jan 21 05:57:10.768057 kernel: audit: type=1103 audit(1768975030.689:877): pid=6579 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:10.768212 kernel: audit: type=1006 audit(1768975030.689:878): pid=6579 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=25 res=1 Jan 21 05:57:10.689000 audit[6579]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffefdb44dd0 a2=3 a3=0 items=0 ppid=1 pid=6579 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:10.828762 kernel: audit: type=1300 audit(1768975030.689:878): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffefdb44dd0 a2=3 a3=0 items=0 ppid=1 pid=6579 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:10.828913 kernel: audit: type=1327 audit(1768975030.689:878): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:10.689000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:10.845497 systemd[1]: Started session-25.scope - Session 25 of User core. Jan 21 05:57:10.860000 audit[6579]: USER_START pid=6579 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:10.867000 audit[6583]: CRED_ACQ pid=6583 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:10.924460 kernel: audit: type=1105 audit(1768975030.860:879): pid=6579 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:10.924770 kernel: audit: type=1103 audit(1768975030.867:880): pid=6583 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:11.193978 sshd[6583]: Connection closed by 10.0.0.1 port 36498 Jan 21 05:57:11.197099 sshd-session[6579]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:11.201000 audit[6579]: USER_END pid=6579 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:11.208454 systemd[1]: sshd@23-10.0.0.61:22-10.0.0.1:36498.service: Deactivated successfully. Jan 21 05:57:11.214951 systemd[1]: session-25.scope: Deactivated successfully. Jan 21 05:57:11.218248 systemd-logind[1620]: Session 25 logged out. Waiting for processes to exit. Jan 21 05:57:11.227291 systemd-logind[1620]: Removed session 25. Jan 21 05:57:11.201000 audit[6579]: CRED_DISP pid=6579 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:11.270953 kernel: audit: type=1106 audit(1768975031.201:881): pid=6579 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:11.271133 kernel: audit: type=1104 audit(1768975031.201:882): pid=6579 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:11.209000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.61:22-10.0.0.1:36498 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:14.212362 kubelet[2982]: E0121 05:57:14.211110 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:57:16.219309 containerd[1639]: time="2026-01-21T05:57:16.219232482Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 21 05:57:16.224449 systemd[1]: Started sshd@24-10.0.0.61:22-10.0.0.1:48110.service - OpenSSH per-connection server daemon (10.0.0.1:48110). Jan 21 05:57:16.249813 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:57:16.249945 kernel: audit: type=1130 audit(1768975036.227:884): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.61:22-10.0.0.1:48110 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:16.227000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.61:22-10.0.0.1:48110 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:16.292155 containerd[1639]: time="2026-01-21T05:57:16.290868685Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:57:16.296027 containerd[1639]: time="2026-01-21T05:57:16.294401022Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 21 05:57:16.296027 containerd[1639]: time="2026-01-21T05:57:16.294515615Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 21 05:57:16.296209 kubelet[2982]: E0121 05:57:16.294874 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 21 05:57:16.296209 kubelet[2982]: E0121 05:57:16.294935 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 21 05:57:16.296209 kubelet[2982]: E0121 05:57:16.295079 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:2d409fc375e643af9c78cec77dff7f0b,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rgkt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-cd95c566-k55c2_calico-system(da0d7215-59d8-4e4a-9787-c2942086bde3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 21 05:57:16.300763 containerd[1639]: time="2026-01-21T05:57:16.300159254Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 21 05:57:16.386529 containerd[1639]: time="2026-01-21T05:57:16.385777326Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:57:16.391298 containerd[1639]: time="2026-01-21T05:57:16.391150320Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 21 05:57:16.391298 containerd[1639]: time="2026-01-21T05:57:16.391258021Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 21 05:57:16.392788 kubelet[2982]: E0121 05:57:16.392011 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 21 05:57:16.392788 kubelet[2982]: E0121 05:57:16.392077 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 21 05:57:16.392788 kubelet[2982]: E0121 05:57:16.392447 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rgkt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-cd95c566-k55c2_calico-system(da0d7215-59d8-4e4a-9787-c2942086bde3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 21 05:57:16.393909 kubelet[2982]: E0121 05:57:16.393680 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-cd95c566-k55c2" podUID="da0d7215-59d8-4e4a-9787-c2942086bde3" Jan 21 05:57:16.396000 audit[6596]: USER_ACCT pid=6596 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.397420 sshd[6596]: Accepted publickey for core from 10.0.0.1 port 48110 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:16.404250 sshd-session[6596]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:16.401000 audit[6596]: CRED_ACQ pid=6596 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.433298 kernel: audit: type=1101 audit(1768975036.396:885): pid=6596 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.433430 kernel: audit: type=1103 audit(1768975036.401:886): pid=6596 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.436825 systemd-logind[1620]: New session 26 of user core. Jan 21 05:57:16.440632 kernel: audit: type=1006 audit(1768975036.401:887): pid=6596 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=26 res=1 Jan 21 05:57:16.440738 kernel: audit: type=1300 audit(1768975036.401:887): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcadc1e9b0 a2=3 a3=0 items=0 ppid=1 pid=6596 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=26 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:16.401000 audit[6596]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcadc1e9b0 a2=3 a3=0 items=0 ppid=1 pid=6596 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=26 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:16.401000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:16.458799 kernel: audit: type=1327 audit(1768975036.401:887): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:16.467315 systemd[1]: Started session-26.scope - Session 26 of User core. Jan 21 05:57:16.480000 audit[6596]: USER_START pid=6596 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.501766 kernel: audit: type=1105 audit(1768975036.480:888): pid=6596 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.501916 kernel: audit: type=1103 audit(1768975036.487:889): pid=6600 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.487000 audit[6600]: CRED_ACQ pid=6600 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.658012 sshd[6600]: Connection closed by 10.0.0.1 port 48110 Jan 21 05:57:16.660922 sshd-session[6596]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:16.664000 audit[6596]: USER_END pid=6596 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.670142 systemd[1]: sshd@24-10.0.0.61:22-10.0.0.1:48110.service: Deactivated successfully. Jan 21 05:57:16.673258 systemd[1]: session-26.scope: Deactivated successfully. Jan 21 05:57:16.674851 systemd-logind[1620]: Session 26 logged out. Waiting for processes to exit. Jan 21 05:57:16.677071 systemd-logind[1620]: Removed session 26. Jan 21 05:57:16.664000 audit[6596]: CRED_DISP pid=6596 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.679740 kernel: audit: type=1106 audit(1768975036.664:890): pid=6596 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.679859 kernel: audit: type=1104 audit(1768975036.664:891): pid=6596 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:16.670000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.61:22-10.0.0.1:48110 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:17.213578 containerd[1639]: time="2026-01-21T05:57:17.213511075Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 21 05:57:17.291911 containerd[1639]: time="2026-01-21T05:57:17.290520272Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:57:17.294125 containerd[1639]: time="2026-01-21T05:57:17.293932172Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 21 05:57:17.295034 containerd[1639]: time="2026-01-21T05:57:17.294117238Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 21 05:57:17.295861 kubelet[2982]: E0121 05:57:17.295396 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 21 05:57:17.295861 kubelet[2982]: E0121 05:57:17.295492 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 21 05:57:17.295992 kubelet[2982]: E0121 05:57:17.295863 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6x2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-m7pjj_calico-system(59842dba-2f25-4ec6-86e9-919aa2afb77a): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 21 05:57:17.297943 kubelet[2982]: E0121 05:57:17.297767 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:57:21.212161 containerd[1639]: time="2026-01-21T05:57:21.211921020Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 21 05:57:21.397750 containerd[1639]: time="2026-01-21T05:57:21.394083617Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:57:21.402756 containerd[1639]: time="2026-01-21T05:57:21.401210367Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 21 05:57:21.402756 containerd[1639]: time="2026-01-21T05:57:21.401352501Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 21 05:57:21.403133 kubelet[2982]: E0121 05:57:21.403045 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:57:21.404734 kubelet[2982]: E0121 05:57:21.403976 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:57:21.404734 kubelet[2982]: E0121 05:57:21.404308 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wclgb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-fcddb958d-sjlxh_calico-apiserver(dbddb9a5-130c-465d-9677-8a7b919a18f8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 21 05:57:21.405753 containerd[1639]: time="2026-01-21T05:57:21.405419118Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 21 05:57:21.406854 kubelet[2982]: E0121 05:57:21.406753 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:57:21.528560 containerd[1639]: time="2026-01-21T05:57:21.528366531Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:57:21.537893 containerd[1639]: time="2026-01-21T05:57:21.536345356Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 21 05:57:21.538532 containerd[1639]: time="2026-01-21T05:57:21.536527939Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 21 05:57:21.539896 kubelet[2982]: E0121 05:57:21.539335 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 21 05:57:21.539896 kubelet[2982]: E0121 05:57:21.539550 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 21 05:57:21.542224 kubelet[2982]: E0121 05:57:21.541325 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8n95t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-7598d7cc9-qv4mq_calico-system(821a7ed1-bcf9-475e-908c-afcf62571645): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 21 05:57:21.545053 kubelet[2982]: E0121 05:57:21.544044 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:57:21.703590 systemd[1]: Started sshd@25-10.0.0.61:22-10.0.0.1:48118.service - OpenSSH per-connection server daemon (10.0.0.1:48118). Jan 21 05:57:21.715055 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:57:21.715849 kernel: audit: type=1130 audit(1768975041.701:893): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.61:22-10.0.0.1:48118 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:21.701000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.61:22-10.0.0.1:48118 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:21.856000 audit[6614]: USER_ACCT pid=6614 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:21.862909 sshd-session[6614]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:21.869227 sshd[6614]: Accepted publickey for core from 10.0.0.1 port 48118 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:21.876375 systemd-logind[1620]: New session 27 of user core. Jan 21 05:57:21.879735 kernel: audit: type=1101 audit(1768975041.856:894): pid=6614 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:21.859000 audit[6614]: CRED_ACQ pid=6614 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:21.909908 kernel: audit: type=1103 audit(1768975041.859:895): pid=6614 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:21.910025 kernel: audit: type=1006 audit(1768975041.859:896): pid=6614 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=27 res=1 Jan 21 05:57:21.911562 kernel: audit: type=1300 audit(1768975041.859:896): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffff4f3eee0 a2=3 a3=0 items=0 ppid=1 pid=6614 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=27 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:21.859000 audit[6614]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffff4f3eee0 a2=3 a3=0 items=0 ppid=1 pid=6614 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=27 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:21.934249 systemd[1]: Started session-27.scope - Session 27 of User core. Jan 21 05:57:21.963775 kernel: audit: type=1327 audit(1768975041.859:896): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:21.859000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:21.966000 audit[6614]: USER_START pid=6614 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:21.977000 audit[6618]: CRED_ACQ pid=6618 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:22.009402 kernel: audit: type=1105 audit(1768975041.966:897): pid=6614 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:22.009724 kernel: audit: type=1103 audit(1768975041.977:898): pid=6618 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:22.492735 containerd[1639]: time="2026-01-21T05:57:22.491748624Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 21 05:57:22.614878 containerd[1639]: time="2026-01-21T05:57:22.611945248Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:57:22.617748 containerd[1639]: time="2026-01-21T05:57:22.617166352Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 21 05:57:22.617748 containerd[1639]: time="2026-01-21T05:57:22.617308207Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 21 05:57:22.618188 kubelet[2982]: E0121 05:57:22.618004 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 21 05:57:22.620001 kubelet[2982]: E0121 05:57:22.618242 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 21 05:57:22.621235 kubelet[2982]: E0121 05:57:22.620811 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgtqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 21 05:57:22.631818 containerd[1639]: time="2026-01-21T05:57:22.628850958Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 21 05:57:22.730832 containerd[1639]: time="2026-01-21T05:57:22.728708753Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:57:22.733099 containerd[1639]: time="2026-01-21T05:57:22.732976936Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 21 05:57:22.733255 containerd[1639]: time="2026-01-21T05:57:22.733132987Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 21 05:57:22.734777 kubelet[2982]: E0121 05:57:22.733384 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 21 05:57:22.734777 kubelet[2982]: E0121 05:57:22.733499 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 21 05:57:22.734777 kubelet[2982]: E0121 05:57:22.733829 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgtqr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-zftdp_calico-system(487c6aa2-04b5-4684-9420-0e4b73799fd3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 21 05:57:22.735332 kubelet[2982]: E0121 05:57:22.735139 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:57:22.747489 sshd[6618]: Connection closed by 10.0.0.1 port 48118 Jan 21 05:57:22.752066 sshd-session[6614]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:22.791000 audit[6614]: USER_END pid=6614 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:22.803153 systemd[1]: sshd@25-10.0.0.61:22-10.0.0.1:48118.service: Deactivated successfully. Jan 21 05:57:22.812348 systemd[1]: session-27.scope: Deactivated successfully. Jan 21 05:57:22.816358 systemd-logind[1620]: Session 27 logged out. Waiting for processes to exit. Jan 21 05:57:22.791000 audit[6614]: CRED_DISP pid=6614 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:22.827936 systemd-logind[1620]: Removed session 27. Jan 21 05:57:22.830170 kernel: audit: type=1106 audit(1768975042.791:899): pid=6614 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:22.830236 kernel: audit: type=1104 audit(1768975042.791:900): pid=6614 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:22.803000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.61:22-10.0.0.1:48118 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:25.214241 containerd[1639]: time="2026-01-21T05:57:25.213558059Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 21 05:57:25.224499 kubelet[2982]: E0121 05:57:25.224297 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:57:25.319128 containerd[1639]: time="2026-01-21T05:57:25.318926545Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 21 05:57:25.323287 containerd[1639]: time="2026-01-21T05:57:25.323218801Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 21 05:57:25.324144 containerd[1639]: time="2026-01-21T05:57:25.323579423Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 21 05:57:25.325206 kubelet[2982]: E0121 05:57:25.325081 2982 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:57:25.325206 kubelet[2982]: E0121 05:57:25.325177 2982 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 21 05:57:25.325778 kubelet[2982]: E0121 05:57:25.325373 2982 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6nlrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-796779bb-6bz64_calico-apiserver(137d495d-bc8e-4cd7-b462-b18325307853): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 21 05:57:25.327215 kubelet[2982]: E0121 05:57:25.327141 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:57:26.213988 kubelet[2982]: E0121 05:57:26.213874 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:57:27.772527 systemd[1]: Started sshd@26-10.0.0.61:22-10.0.0.1:45348.service - OpenSSH per-connection server daemon (10.0.0.1:45348). Jan 21 05:57:27.772000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.61:22-10.0.0.1:45348 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:27.776909 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:57:27.777020 kernel: audit: type=1130 audit(1768975047.772:902): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.61:22-10.0.0.1:45348 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:27.869000 audit[6641]: USER_ACCT pid=6641 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:27.872066 sshd[6641]: Accepted publickey for core from 10.0.0.1 port 45348 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:27.878986 sshd-session[6641]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:27.886893 kernel: audit: type=1101 audit(1768975047.869:903): pid=6641 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:27.876000 audit[6641]: CRED_ACQ pid=6641 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:27.892018 systemd-logind[1620]: New session 28 of user core. Jan 21 05:57:27.900804 kernel: audit: type=1103 audit(1768975047.876:904): pid=6641 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:27.876000 audit[6641]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc7137efb0 a2=3 a3=0 items=0 ppid=1 pid=6641 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=28 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:27.924738 kernel: audit: type=1006 audit(1768975047.876:905): pid=6641 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=28 res=1 Jan 21 05:57:27.924887 kernel: audit: type=1300 audit(1768975047.876:905): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc7137efb0 a2=3 a3=0 items=0 ppid=1 pid=6641 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=28 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:27.924963 kernel: audit: type=1327 audit(1768975047.876:905): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:27.876000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:27.926105 systemd[1]: Started session-28.scope - Session 28 of User core. Jan 21 05:57:27.933000 audit[6641]: USER_START pid=6641 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:27.938000 audit[6645]: CRED_ACQ pid=6645 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:27.984362 kernel: audit: type=1105 audit(1768975047.933:906): pid=6641 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:27.985019 kernel: audit: type=1103 audit(1768975047.938:907): pid=6645 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:28.166815 sshd[6645]: Connection closed by 10.0.0.1 port 45348 Jan 21 05:57:28.165171 sshd-session[6641]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:28.167000 audit[6641]: USER_END pid=6641 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:28.167000 audit[6641]: CRED_DISP pid=6641 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:28.204040 kernel: audit: type=1106 audit(1768975048.167:908): pid=6641 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:28.204149 kernel: audit: type=1104 audit(1768975048.167:909): pid=6641 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:28.213036 systemd[1]: sshd@26-10.0.0.61:22-10.0.0.1:45348.service: Deactivated successfully. Jan 21 05:57:28.212000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.61:22-10.0.0.1:45348 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:28.217295 kubelet[2982]: E0121 05:57:28.217174 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-cd95c566-k55c2" podUID="da0d7215-59d8-4e4a-9787-c2942086bde3" Jan 21 05:57:28.219530 systemd[1]: session-28.scope: Deactivated successfully. Jan 21 05:57:28.224149 systemd-logind[1620]: Session 28 logged out. Waiting for processes to exit. Jan 21 05:57:28.234035 systemd[1]: Started sshd@27-10.0.0.61:22-10.0.0.1:45350.service - OpenSSH per-connection server daemon (10.0.0.1:45350). Jan 21 05:57:28.234000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.61:22-10.0.0.1:45350 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:28.239370 systemd-logind[1620]: Removed session 28. Jan 21 05:57:28.334000 audit[6659]: USER_ACCT pid=6659 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:28.335817 sshd[6659]: Accepted publickey for core from 10.0.0.1 port 45350 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:28.336000 audit[6659]: CRED_ACQ pid=6659 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:28.336000 audit[6659]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc62500240 a2=3 a3=0 items=0 ppid=1 pid=6659 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=29 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:28.336000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:28.340201 sshd-session[6659]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:28.369424 systemd-logind[1620]: New session 29 of user core. Jan 21 05:57:28.380585 systemd[1]: Started session-29.scope - Session 29 of User core. Jan 21 05:57:28.385000 audit[6659]: USER_START pid=6659 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:28.389000 audit[6663]: CRED_ACQ pid=6663 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:29.116194 sshd[6663]: Connection closed by 10.0.0.1 port 45350 Jan 21 05:57:29.118249 sshd-session[6659]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:29.122000 audit[6659]: USER_END pid=6659 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:29.122000 audit[6659]: CRED_DISP pid=6659 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:29.136915 systemd[1]: sshd@27-10.0.0.61:22-10.0.0.1:45350.service: Deactivated successfully. Jan 21 05:57:29.136000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.61:22-10.0.0.1:45350 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:29.142217 systemd[1]: session-29.scope: Deactivated successfully. Jan 21 05:57:29.145799 systemd-logind[1620]: Session 29 logged out. Waiting for processes to exit. Jan 21 05:57:29.156153 systemd-logind[1620]: Removed session 29. Jan 21 05:57:29.161351 systemd[1]: Started sshd@28-10.0.0.61:22-10.0.0.1:45356.service - OpenSSH per-connection server daemon (10.0.0.1:45356). Jan 21 05:57:29.160000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.61:22-10.0.0.1:45356 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:29.332000 audit[6675]: USER_ACCT pid=6675 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:29.334000 audit[6675]: CRED_ACQ pid=6675 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:29.335000 audit[6675]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff28cafbe0 a2=3 a3=0 items=0 ppid=1 pid=6675 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=30 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:29.335000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:29.353478 sshd[6675]: Accepted publickey for core from 10.0.0.1 port 45356 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:29.337186 sshd-session[6675]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:29.356807 systemd-logind[1620]: New session 30 of user core. Jan 21 05:57:29.401903 systemd[1]: Started session-30.scope - Session 30 of User core. Jan 21 05:57:29.407000 audit[6675]: USER_START pid=6675 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:29.411000 audit[6679]: CRED_ACQ pid=6679 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:30.211747 kubelet[2982]: E0121 05:57:30.211212 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:57:30.325000 audit[6718]: NETFILTER_CFG table=filter:140 family=2 entries=26 op=nft_register_rule pid=6718 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:57:30.325000 audit[6718]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffccdf8d8f0 a2=0 a3=7ffccdf8d8dc items=0 ppid=3144 pid=6718 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:30.325000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:57:30.335000 audit[6718]: NETFILTER_CFG table=nat:141 family=2 entries=20 op=nft_register_rule pid=6718 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:57:30.335000 audit[6718]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffccdf8d8f0 a2=0 a3=0 items=0 ppid=3144 pid=6718 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:30.335000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:57:30.365265 sshd[6679]: Connection closed by 10.0.0.1 port 45356 Jan 21 05:57:30.367181 sshd-session[6675]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:30.370000 audit[6675]: USER_END pid=6675 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:30.370000 audit[6675]: CRED_DISP pid=6675 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:30.377251 systemd[1]: sshd@28-10.0.0.61:22-10.0.0.1:45356.service: Deactivated successfully. Jan 21 05:57:30.377000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.61:22-10.0.0.1:45356 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:30.382226 systemd[1]: session-30.scope: Deactivated successfully. Jan 21 05:57:30.383000 audit[6721]: NETFILTER_CFG table=filter:142 family=2 entries=38 op=nft_register_rule pid=6721 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:57:30.383000 audit[6721]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffe898e9030 a2=0 a3=7ffe898e901c items=0 ppid=3144 pid=6721 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:30.383000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:57:30.387086 systemd-logind[1620]: Session 30 logged out. Waiting for processes to exit. Jan 21 05:57:30.389000 audit[6721]: NETFILTER_CFG table=nat:143 family=2 entries=20 op=nft_register_rule pid=6721 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:57:30.389000 audit[6721]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffe898e9030 a2=0 a3=0 items=0 ppid=3144 pid=6721 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:30.389000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:57:30.395245 systemd[1]: Started sshd@29-10.0.0.61:22-10.0.0.1:45370.service - OpenSSH per-connection server daemon (10.0.0.1:45370). Jan 21 05:57:30.394000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.61:22-10.0.0.1:45370 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:30.396961 systemd-logind[1620]: Removed session 30. Jan 21 05:57:30.505000 audit[6725]: USER_ACCT pid=6725 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:30.506881 sshd[6725]: Accepted publickey for core from 10.0.0.1 port 45370 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:30.507000 audit[6725]: CRED_ACQ pid=6725 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:30.507000 audit[6725]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdae979100 a2=3 a3=0 items=0 ppid=1 pid=6725 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=31 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:30.507000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:30.509744 sshd-session[6725]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:30.519401 systemd-logind[1620]: New session 31 of user core. Jan 21 05:57:30.529187 systemd[1]: Started session-31.scope - Session 31 of User core. Jan 21 05:57:30.534000 audit[6725]: USER_START pid=6725 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:30.539000 audit[6729]: CRED_ACQ pid=6729 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:31.006081 sshd[6729]: Connection closed by 10.0.0.1 port 45370 Jan 21 05:57:31.009406 sshd-session[6725]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:31.011000 audit[6725]: USER_END pid=6725 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:31.011000 audit[6725]: CRED_DISP pid=6725 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:31.021506 systemd[1]: sshd@29-10.0.0.61:22-10.0.0.1:45370.service: Deactivated successfully. Jan 21 05:57:31.021000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.61:22-10.0.0.1:45370 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:31.028223 systemd[1]: session-31.scope: Deactivated successfully. Jan 21 05:57:31.032170 systemd-logind[1620]: Session 31 logged out. Waiting for processes to exit. Jan 21 05:57:31.040770 systemd[1]: Started sshd@30-10.0.0.61:22-10.0.0.1:45382.service - OpenSSH per-connection server daemon (10.0.0.1:45382). Jan 21 05:57:31.040000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@30-10.0.0.61:22-10.0.0.1:45382 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:31.041734 systemd-logind[1620]: Removed session 31. Jan 21 05:57:31.130000 audit[6742]: USER_ACCT pid=6742 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:31.131615 sshd[6742]: Accepted publickey for core from 10.0.0.1 port 45382 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:31.133000 audit[6742]: CRED_ACQ pid=6742 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:31.133000 audit[6742]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd4c2c66d0 a2=3 a3=0 items=0 ppid=1 pid=6742 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=32 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:31.133000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:31.136239 sshd-session[6742]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:31.153523 systemd-logind[1620]: New session 32 of user core. Jan 21 05:57:31.163472 systemd[1]: Started session-32.scope - Session 32 of User core. Jan 21 05:57:31.167000 audit[6742]: USER_START pid=6742 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:31.170000 audit[6746]: CRED_ACQ pid=6746 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:31.311985 sshd[6746]: Connection closed by 10.0.0.1 port 45382 Jan 21 05:57:31.313936 sshd-session[6742]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:31.316000 audit[6742]: USER_END pid=6742 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:31.316000 audit[6742]: CRED_DISP pid=6742 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:31.321973 systemd[1]: sshd@30-10.0.0.61:22-10.0.0.1:45382.service: Deactivated successfully. Jan 21 05:57:31.321000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@30-10.0.0.61:22-10.0.0.1:45382 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:31.325562 systemd[1]: session-32.scope: Deactivated successfully. Jan 21 05:57:31.329475 systemd-logind[1620]: Session 32 logged out. Waiting for processes to exit. Jan 21 05:57:31.331589 systemd-logind[1620]: Removed session 32. Jan 21 05:57:33.211507 kubelet[2982]: E0121 05:57:33.211426 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:57:35.213376 kubelet[2982]: E0121 05:57:35.213277 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3" Jan 21 05:57:36.212186 kubelet[2982]: E0121 05:57:36.211210 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-fcddb958d-sjlxh" podUID="dbddb9a5-130c-465d-9677-8a7b919a18f8" Jan 21 05:57:36.340221 systemd[1]: Started sshd@31-10.0.0.61:22-10.0.0.1:60816.service - OpenSSH per-connection server daemon (10.0.0.1:60816). Jan 21 05:57:36.339000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@31-10.0.0.61:22-10.0.0.1:60816 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:36.356871 kernel: kauditd_printk_skb: 57 callbacks suppressed Jan 21 05:57:36.357080 kernel: audit: type=1130 audit(1768975056.339:951): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@31-10.0.0.61:22-10.0.0.1:60816 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:36.450000 audit[6761]: USER_ACCT pid=6761 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.451982 sshd[6761]: Accepted publickey for core from 10.0.0.1 port 60816 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:36.454308 sshd-session[6761]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:36.461949 systemd-logind[1620]: New session 33 of user core. Jan 21 05:57:36.452000 audit[6761]: CRED_ACQ pid=6761 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.480095 kernel: audit: type=1101 audit(1768975056.450:952): pid=6761 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.480419 kernel: audit: type=1103 audit(1768975056.452:953): pid=6761 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.480449 kernel: audit: type=1006 audit(1768975056.452:954): pid=6761 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=33 res=1 Jan 21 05:57:36.488536 kernel: audit: type=1300 audit(1768975056.452:954): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc05e72970 a2=3 a3=0 items=0 ppid=1 pid=6761 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=33 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:36.452000 audit[6761]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc05e72970 a2=3 a3=0 items=0 ppid=1 pid=6761 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=33 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:36.501771 kernel: audit: type=1327 audit(1768975056.452:954): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:36.452000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:36.509418 systemd[1]: Started session-33.scope - Session 33 of User core. Jan 21 05:57:36.514000 audit[6761]: USER_START pid=6761 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.518000 audit[6765]: CRED_ACQ pid=6765 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.547726 kernel: audit: type=1105 audit(1768975056.514:955): pid=6761 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.547881 kernel: audit: type=1103 audit(1768975056.518:956): pid=6765 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.664166 sshd[6765]: Connection closed by 10.0.0.1 port 60816 Jan 21 05:57:36.664778 sshd-session[6761]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:36.666000 audit[6761]: USER_END pid=6761 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.672167 systemd[1]: sshd@31-10.0.0.61:22-10.0.0.1:60816.service: Deactivated successfully. Jan 21 05:57:36.675782 systemd[1]: session-33.scope: Deactivated successfully. Jan 21 05:57:36.677789 systemd-logind[1620]: Session 33 logged out. Waiting for processes to exit. Jan 21 05:57:36.680430 systemd-logind[1620]: Removed session 33. Jan 21 05:57:36.666000 audit[6761]: CRED_DISP pid=6761 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.693349 kernel: audit: type=1106 audit(1768975056.666:957): pid=6761 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.693500 kernel: audit: type=1104 audit(1768975056.666:958): pid=6761 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:36.672000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@31-10.0.0.61:22-10.0.0.1:60816 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:37.208796 kubelet[2982]: E0121 05:57:37.208444 2982 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 21 05:57:37.209908 kubelet[2982]: E0121 05:57:37.209826 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-f2fds" podUID="dac6efad-1039-4242-b1c8-782ac61e5470" Jan 21 05:57:37.434000 audit[6778]: NETFILTER_CFG table=filter:144 family=2 entries=26 op=nft_register_rule pid=6778 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:57:37.434000 audit[6778]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffd98057cd0 a2=0 a3=7ffd98057cbc items=0 ppid=3144 pid=6778 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:37.434000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:57:37.452000 audit[6778]: NETFILTER_CFG table=nat:145 family=2 entries=104 op=nft_register_chain pid=6778 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 21 05:57:37.452000 audit[6778]: SYSCALL arch=c000003e syscall=46 success=yes exit=48684 a0=3 a1=7ffd98057cd0 a2=0 a3=7ffd98057cbc items=0 ppid=3144 pid=6778 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:37.452000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 21 05:57:39.211456 kubelet[2982]: E0121 05:57:39.211309 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-796779bb-6bz64" podUID="137d495d-bc8e-4cd7-b462-b18325307853" Jan 21 05:57:41.688944 systemd[1]: Started sshd@32-10.0.0.61:22-10.0.0.1:60830.service - OpenSSH per-connection server daemon (10.0.0.1:60830). Jan 21 05:57:41.696145 kernel: kauditd_printk_skb: 7 callbacks suppressed Jan 21 05:57:41.696243 kernel: audit: type=1130 audit(1768975061.688:962): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@32-10.0.0.61:22-10.0.0.1:60830 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:41.688000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@32-10.0.0.61:22-10.0.0.1:60830 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:41.803000 audit[6780]: USER_ACCT pid=6780 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:41.805107 sshd[6780]: Accepted publickey for core from 10.0.0.1 port 60830 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:41.809149 sshd-session[6780]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:41.806000 audit[6780]: CRED_ACQ pid=6780 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:41.819208 systemd-logind[1620]: New session 34 of user core. Jan 21 05:57:41.829985 kernel: audit: type=1101 audit(1768975061.803:963): pid=6780 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:41.830132 kernel: audit: type=1103 audit(1768975061.806:964): pid=6780 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:41.830176 kernel: audit: type=1006 audit(1768975061.806:965): pid=6780 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=34 res=1 Jan 21 05:57:41.853962 kernel: audit: type=1300 audit(1768975061.806:965): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe91fd92f0 a2=3 a3=0 items=0 ppid=1 pid=6780 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=34 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:41.806000 audit[6780]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe91fd92f0 a2=3 a3=0 items=0 ppid=1 pid=6780 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=34 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:41.870030 kernel: audit: type=1327 audit(1768975061.806:965): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:41.806000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:41.883035 systemd[1]: Started session-34.scope - Session 34 of User core. Jan 21 05:57:41.889000 audit[6780]: USER_START pid=6780 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:41.889000 audit[6784]: CRED_ACQ pid=6784 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:41.924811 kernel: audit: type=1105 audit(1768975061.889:966): pid=6780 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:41.925037 kernel: audit: type=1103 audit(1768975061.889:967): pid=6784 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:42.092083 sshd[6784]: Connection closed by 10.0.0.1 port 60830 Jan 21 05:57:42.092771 sshd-session[6780]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:42.095000 audit[6780]: USER_END pid=6780 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:42.101390 systemd[1]: sshd@32-10.0.0.61:22-10.0.0.1:60830.service: Deactivated successfully. Jan 21 05:57:42.115358 kernel: audit: type=1106 audit(1768975062.095:968): pid=6780 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:42.115457 kernel: audit: type=1104 audit(1768975062.095:969): pid=6780 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:42.095000 audit[6780]: CRED_DISP pid=6780 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:42.112157 systemd[1]: session-34.scope: Deactivated successfully. Jan 21 05:57:42.124726 systemd-logind[1620]: Session 34 logged out. Waiting for processes to exit. Jan 21 05:57:42.126841 systemd-logind[1620]: Removed session 34. Jan 21 05:57:42.101000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@32-10.0.0.61:22-10.0.0.1:60830 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:42.212003 kubelet[2982]: E0121 05:57:42.211767 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-cd95c566-k55c2" podUID="da0d7215-59d8-4e4a-9787-c2942086bde3" Jan 21 05:57:43.214084 kubelet[2982]: E0121 05:57:43.213027 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-m7pjj" podUID="59842dba-2f25-4ec6-86e9-919aa2afb77a" Jan 21 05:57:46.216154 kubelet[2982]: E0121 05:57:46.216093 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7598d7cc9-qv4mq" podUID="821a7ed1-bcf9-475e-908c-afcf62571645" Jan 21 05:57:47.123006 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 21 05:57:47.123327 kernel: audit: type=1130 audit(1768975067.114:971): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@33-10.0.0.61:22-10.0.0.1:55400 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:47.114000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@33-10.0.0.61:22-10.0.0.1:55400 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:47.115104 systemd[1]: Started sshd@33-10.0.0.61:22-10.0.0.1:55400.service - OpenSSH per-connection server daemon (10.0.0.1:55400). Jan 21 05:57:47.250000 audit[6798]: USER_ACCT pid=6798 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.266918 kernel: audit: type=1101 audit(1768975067.250:972): pid=6798 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.260516 sshd-session[6798]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 21 05:57:47.267583 sshd[6798]: Accepted publickey for core from 10.0.0.1 port 55400 ssh2: RSA SHA256:cfT5WYkUgCbS6oPyID9sampNshbVcF4fd8PmhQlgHRE Jan 21 05:57:47.257000 audit[6798]: CRED_ACQ pid=6798 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.290067 kernel: audit: type=1103 audit(1768975067.257:973): pid=6798 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.290177 kernel: audit: type=1006 audit(1768975067.257:974): pid=6798 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=35 res=1 Jan 21 05:57:47.288933 systemd-logind[1620]: New session 35 of user core. Jan 21 05:57:47.257000 audit[6798]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdb2720e80 a2=3 a3=0 items=0 ppid=1 pid=6798 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=35 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:47.257000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:47.311479 kernel: audit: type=1300 audit(1768975067.257:974): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdb2720e80 a2=3 a3=0 items=0 ppid=1 pid=6798 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=35 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 21 05:57:47.311571 kernel: audit: type=1327 audit(1768975067.257:974): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 21 05:57:47.315600 systemd[1]: Started session-35.scope - Session 35 of User core. Jan 21 05:57:47.333000 audit[6798]: USER_START pid=6798 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.357790 kernel: audit: type=1105 audit(1768975067.333:975): pid=6798 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.341000 audit[6802]: CRED_ACQ pid=6802 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.376777 kernel: audit: type=1103 audit(1768975067.341:976): pid=6802 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.551113 sshd[6802]: Connection closed by 10.0.0.1 port 55400 Jan 21 05:57:47.551971 sshd-session[6798]: pam_unix(sshd:session): session closed for user core Jan 21 05:57:47.557000 audit[6798]: USER_END pid=6798 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.566219 systemd[1]: sshd@33-10.0.0.61:22-10.0.0.1:55400.service: Deactivated successfully. Jan 21 05:57:47.577375 systemd[1]: session-35.scope: Deactivated successfully. Jan 21 05:57:47.577722 kernel: audit: type=1106 audit(1768975067.557:977): pid=6798 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.557000 audit[6798]: CRED_DISP pid=6798 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.583410 systemd-logind[1620]: Session 35 logged out. Waiting for processes to exit. Jan 21 05:57:47.586244 systemd-logind[1620]: Removed session 35. Jan 21 05:57:47.592958 kernel: audit: type=1104 audit(1768975067.557:978): pid=6798 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 21 05:57:47.567000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@33-10.0.0.61:22-10.0.0.1:55400 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 21 05:57:48.217305 kubelet[2982]: E0121 05:57:48.217164 2982 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-zftdp" podUID="487c6aa2-04b5-4684-9420-0e4b73799fd3"