Jan 24 11:36:17.349559 kernel: Linux version 6.12.66-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.1_p20250801 p4) 14.3.1 20250801, GNU ld (Gentoo 2.45 p3) 2.45.0) #1 SMP PREEMPT_DYNAMIC Sat Jan 24 09:07:34 -00 2026 Jan 24 11:36:17.349598 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=7953d3c7acaad6ee79638a10c67ea9f0b3a8597919989b6fbf2f9a1742d4ba63 Jan 24 11:36:17.349614 kernel: BIOS-provided physical RAM map: Jan 24 11:36:17.349630 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Jan 24 11:36:17.349638 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Jan 24 11:36:17.349647 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Jan 24 11:36:17.349659 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Jan 24 11:36:17.349671 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Jan 24 11:36:17.349847 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Jan 24 11:36:17.349860 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Jan 24 11:36:17.349868 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Jan 24 11:36:17.349883 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Jan 24 11:36:17.349895 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Jan 24 11:36:17.349906 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Jan 24 11:36:17.349916 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Jan 24 11:36:17.349926 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Jan 24 11:36:17.349997 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Jan 24 11:36:17.350011 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Jan 24 11:36:17.350020 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Jan 24 11:36:17.350030 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Jan 24 11:36:17.350039 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Jan 24 11:36:17.350051 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Jan 24 11:36:17.350063 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Jan 24 11:36:17.350072 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 24 11:36:17.350082 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Jan 24 11:36:17.350093 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 24 11:36:17.350181 kernel: NX (Execute Disable) protection: active Jan 24 11:36:17.350192 kernel: APIC: Static calls initialized Jan 24 11:36:17.350202 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Jan 24 11:36:17.350212 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Jan 24 11:36:17.350222 kernel: extended physical RAM map: Jan 24 11:36:17.350234 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Jan 24 11:36:17.350245 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Jan 24 11:36:17.350254 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Jan 24 11:36:17.350264 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Jan 24 11:36:17.350276 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Jan 24 11:36:17.350288 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Jan 24 11:36:17.350302 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Jan 24 11:36:17.350312 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Jan 24 11:36:17.350324 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Jan 24 11:36:17.350341 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Jan 24 11:36:17.350354 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Jan 24 11:36:17.350366 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Jan 24 11:36:17.350379 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Jan 24 11:36:17.350390 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Jan 24 11:36:17.350400 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Jan 24 11:36:17.350412 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Jan 24 11:36:17.350425 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Jan 24 11:36:17.350435 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Jan 24 11:36:17.350445 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Jan 24 11:36:17.350461 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Jan 24 11:36:17.350473 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Jan 24 11:36:17.350484 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Jan 24 11:36:17.350494 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Jan 24 11:36:17.350507 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Jan 24 11:36:17.350519 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 24 11:36:17.350533 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Jan 24 11:36:17.350544 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 24 11:36:17.350617 kernel: efi: EFI v2.7 by EDK II Jan 24 11:36:17.350633 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Jan 24 11:36:17.350701 kernel: random: crng init done Jan 24 11:36:17.350836 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Jan 24 11:36:17.350908 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Jan 24 11:36:17.350922 kernel: secureboot: Secure boot disabled Jan 24 11:36:17.350932 kernel: SMBIOS 2.8 present. Jan 24 11:36:17.350942 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Jan 24 11:36:17.350953 kernel: DMI: Memory slots populated: 1/1 Jan 24 11:36:17.350965 kernel: Hypervisor detected: KVM Jan 24 11:36:17.350976 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Jan 24 11:36:17.350989 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jan 24 11:36:17.351000 kernel: kvm-clock: using sched offset of 21184271616 cycles Jan 24 11:36:17.351011 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jan 24 11:36:17.351028 kernel: tsc: Detected 2445.426 MHz processor Jan 24 11:36:17.351042 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 24 11:36:17.351055 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 24 11:36:17.351066 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Jan 24 11:36:17.351076 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Jan 24 11:36:17.351087 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 24 11:36:17.351169 kernel: Using GB pages for direct mapping Jan 24 11:36:17.351186 kernel: ACPI: Early table checksum verification disabled Jan 24 11:36:17.351199 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Jan 24 11:36:17.351212 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Jan 24 11:36:17.351224 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:36:17.351237 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:36:17.351249 kernel: ACPI: FACS 0x000000009CBDD000 000040 Jan 24 11:36:17.351262 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:36:17.351278 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:36:17.351291 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:36:17.351303 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:36:17.351315 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Jan 24 11:36:17.351327 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Jan 24 11:36:17.351339 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Jan 24 11:36:17.351352 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Jan 24 11:36:17.351368 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Jan 24 11:36:17.351380 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Jan 24 11:36:17.351392 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Jan 24 11:36:17.351404 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Jan 24 11:36:17.351416 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Jan 24 11:36:17.351428 kernel: No NUMA configuration found Jan 24 11:36:17.351440 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Jan 24 11:36:17.351456 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Jan 24 11:36:17.351468 kernel: Zone ranges: Jan 24 11:36:17.351480 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 24 11:36:17.351493 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Jan 24 11:36:17.351505 kernel: Normal empty Jan 24 11:36:17.351517 kernel: Device empty Jan 24 11:36:17.351529 kernel: Movable zone start for each node Jan 24 11:36:17.351541 kernel: Early memory node ranges Jan 24 11:36:17.351557 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Jan 24 11:36:17.351629 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Jan 24 11:36:17.351645 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Jan 24 11:36:17.351656 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Jan 24 11:36:17.351666 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Jan 24 11:36:17.351677 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Jan 24 11:36:17.351691 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Jan 24 11:36:17.351709 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Jan 24 11:36:17.351910 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Jan 24 11:36:17.351977 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 24 11:36:17.352003 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Jan 24 11:36:17.352080 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Jan 24 11:36:17.352342 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 24 11:36:17.352403 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Jan 24 11:36:17.352416 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Jan 24 11:36:17.352534 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Jan 24 11:36:17.352644 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Jan 24 11:36:17.352665 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Jan 24 11:36:17.352676 kernel: ACPI: PM-Timer IO Port: 0x608 Jan 24 11:36:17.352687 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jan 24 11:36:17.352706 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jan 24 11:36:17.352877 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jan 24 11:36:17.352892 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jan 24 11:36:17.352903 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jan 24 11:36:17.352914 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jan 24 11:36:17.352927 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jan 24 11:36:17.352940 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 24 11:36:17.352957 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jan 24 11:36:17.352968 kernel: TSC deadline timer available Jan 24 11:36:17.352983 kernel: CPU topo: Max. logical packages: 1 Jan 24 11:36:17.352995 kernel: CPU topo: Max. logical dies: 1 Jan 24 11:36:17.353006 kernel: CPU topo: Max. dies per package: 1 Jan 24 11:36:17.353074 kernel: CPU topo: Max. threads per core: 1 Jan 24 11:36:17.353091 kernel: CPU topo: Num. cores per package: 4 Jan 24 11:36:17.353172 kernel: CPU topo: Num. threads per package: 4 Jan 24 11:36:17.353190 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Jan 24 11:36:17.353204 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jan 24 11:36:17.353216 kernel: kvm-guest: KVM setup pv remote TLB flush Jan 24 11:36:17.353227 kernel: kvm-guest: setup PV sched yield Jan 24 11:36:17.353238 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Jan 24 11:36:17.353252 kernel: Booting paravirtualized kernel on KVM Jan 24 11:36:17.353264 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 24 11:36:17.353281 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Jan 24 11:36:17.353294 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Jan 24 11:36:17.353308 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Jan 24 11:36:17.353319 kernel: pcpu-alloc: [0] 0 1 2 3 Jan 24 11:36:17.353330 kernel: kvm-guest: PV spinlocks enabled Jan 24 11:36:17.353343 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jan 24 11:36:17.353455 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=7953d3c7acaad6ee79638a10c67ea9f0b3a8597919989b6fbf2f9a1742d4ba63 Jan 24 11:36:17.353618 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 24 11:36:17.353633 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 24 11:36:17.353703 kernel: Fallback order for Node 0: 0 Jan 24 11:36:17.354064 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Jan 24 11:36:17.354076 kernel: Policy zone: DMA32 Jan 24 11:36:17.354086 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 24 11:36:17.354169 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Jan 24 11:36:17.354190 kernel: ftrace: allocating 40128 entries in 157 pages Jan 24 11:36:17.354205 kernel: ftrace: allocated 157 pages with 5 groups Jan 24 11:36:17.354216 kernel: Dynamic Preempt: voluntary Jan 24 11:36:17.354227 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 24 11:36:17.354239 kernel: rcu: RCU event tracing is enabled. Jan 24 11:36:17.354254 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Jan 24 11:36:17.354265 kernel: Trampoline variant of Tasks RCU enabled. Jan 24 11:36:17.354454 kernel: Rude variant of Tasks RCU enabled. Jan 24 11:36:17.354524 kernel: Tracing variant of Tasks RCU enabled. Jan 24 11:36:17.354644 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 24 11:36:17.354708 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Jan 24 11:36:17.354944 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 24 11:36:17.354963 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 24 11:36:17.354977 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 24 11:36:17.354996 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Jan 24 11:36:17.355008 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 24 11:36:17.355020 kernel: Console: colour dummy device 80x25 Jan 24 11:36:17.355035 kernel: printk: legacy console [ttyS0] enabled Jan 24 11:36:17.355049 kernel: ACPI: Core revision 20240827 Jan 24 11:36:17.355061 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jan 24 11:36:17.355072 kernel: APIC: Switch to symmetric I/O mode setup Jan 24 11:36:17.355087 kernel: x2apic enabled Jan 24 11:36:17.355173 kernel: APIC: Switched APIC routing to: physical x2apic Jan 24 11:36:17.355189 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jan 24 11:36:17.355201 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jan 24 11:36:17.355213 kernel: kvm-guest: setup PV IPIs Jan 24 11:36:17.355274 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 24 11:36:17.355288 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Jan 24 11:36:17.355523 kernel: Calibrating delay loop (skipped) preset value.. 4890.85 BogoMIPS (lpj=2445426) Jan 24 11:36:17.355602 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jan 24 11:36:17.355671 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jan 24 11:36:17.355685 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jan 24 11:36:17.355697 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 24 11:36:17.356084 kernel: Spectre V2 : Mitigation: Retpolines Jan 24 11:36:17.356173 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Jan 24 11:36:17.356193 kernel: Speculative Store Bypass: Vulnerable Jan 24 11:36:17.356205 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jan 24 11:36:17.356218 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jan 24 11:36:17.356336 kernel: active return thunk: srso_alias_return_thunk Jan 24 11:36:17.356355 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jan 24 11:36:17.356368 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Jan 24 11:36:17.356380 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Jan 24 11:36:17.356398 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 24 11:36:17.356411 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 24 11:36:17.356423 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 24 11:36:17.356435 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 24 11:36:17.356514 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jan 24 11:36:17.356594 kernel: Freeing SMP alternatives memory: 32K Jan 24 11:36:17.356606 kernel: pid_max: default: 32768 minimum: 301 Jan 24 11:36:17.356623 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Jan 24 11:36:17.356637 kernel: landlock: Up and running. Jan 24 11:36:17.356651 kernel: SELinux: Initializing. Jan 24 11:36:17.356662 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 24 11:36:17.356673 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 24 11:36:17.356686 kernel: smpboot: CPU0: AMD EPYC 7763 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Jan 24 11:36:17.356698 kernel: Performance Events: PMU not available due to virtualization, using software events only. Jan 24 11:36:17.356937 kernel: signal: max sigframe size: 1776 Jan 24 11:36:17.356956 kernel: rcu: Hierarchical SRCU implementation. Jan 24 11:36:17.356969 kernel: rcu: Max phase no-delay instances is 400. Jan 24 11:36:17.356981 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Jan 24 11:36:17.356992 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jan 24 11:36:17.357004 kernel: smp: Bringing up secondary CPUs ... Jan 24 11:36:17.357017 kernel: smpboot: x86: Booting SMP configuration: Jan 24 11:36:17.357174 kernel: .... node #0, CPUs: #1 #2 #3 Jan 24 11:36:17.357190 kernel: smp: Brought up 1 node, 4 CPUs Jan 24 11:36:17.357203 kernel: smpboot: Total of 4 processors activated (19563.40 BogoMIPS) Jan 24 11:36:17.357287 kernel: Memory: 2439048K/2565800K available (14336K kernel code, 2445K rwdata, 31644K rodata, 15536K init, 2500K bss, 120812K reserved, 0K cma-reserved) Jan 24 11:36:17.357303 kernel: devtmpfs: initialized Jan 24 11:36:17.357316 kernel: x86/mm: Memory block size: 128MB Jan 24 11:36:17.357329 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Jan 24 11:36:17.357348 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Jan 24 11:36:17.357361 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Jan 24 11:36:17.357374 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Jan 24 11:36:17.357388 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Jan 24 11:36:17.357402 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Jan 24 11:36:17.357413 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 24 11:36:17.357424 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Jan 24 11:36:17.357440 kernel: pinctrl core: initialized pinctrl subsystem Jan 24 11:36:17.357454 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 24 11:36:17.357535 kernel: audit: initializing netlink subsys (disabled) Jan 24 11:36:17.357550 kernel: audit: type=2000 audit(1769254543.138:1): state=initialized audit_enabled=0 res=1 Jan 24 11:36:17.357565 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 24 11:36:17.357636 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 24 11:36:17.357708 kernel: cpuidle: using governor menu Jan 24 11:36:17.358231 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 24 11:36:17.358349 kernel: dca service started, version 1.12.1 Jan 24 11:36:17.358364 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Jan 24 11:36:17.358419 kernel: PCI: Using configuration type 1 for base access Jan 24 11:36:17.358473 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 24 11:36:17.358487 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 24 11:36:17.358549 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jan 24 11:36:17.358668 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 24 11:36:17.358902 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 24 11:36:17.358918 kernel: ACPI: Added _OSI(Module Device) Jan 24 11:36:17.358931 kernel: ACPI: Added _OSI(Processor Device) Jan 24 11:36:17.358944 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 24 11:36:17.358956 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 24 11:36:17.358969 kernel: ACPI: Interpreter enabled Jan 24 11:36:17.358987 kernel: ACPI: PM: (supports S0 S3 S5) Jan 24 11:36:17.358999 kernel: ACPI: Using IOAPIC for interrupt routing Jan 24 11:36:17.359012 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 24 11:36:17.359072 kernel: PCI: Using E820 reservations for host bridge windows Jan 24 11:36:17.359085 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Jan 24 11:36:17.359163 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 24 11:36:17.361678 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 24 11:36:17.365224 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Jan 24 11:36:17.366488 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Jan 24 11:36:17.366509 kernel: PCI host bridge to bus 0000:00 Jan 24 11:36:17.367319 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 24 11:36:17.367865 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jan 24 11:36:17.368229 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 24 11:36:17.369032 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Jan 24 11:36:17.369474 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Jan 24 11:36:17.369870 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Jan 24 11:36:17.371182 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 24 11:36:17.371518 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Jan 24 11:36:17.372042 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Jan 24 11:36:17.372867 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Jan 24 11:36:17.373244 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Jan 24 11:36:17.373533 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Jan 24 11:36:17.373964 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 24 11:36:17.374397 kernel: pci 0000:00:01.0: pci_fixup_video+0x0/0x100 took 16601 usecs Jan 24 11:36:17.374712 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Jan 24 11:36:17.375219 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Jan 24 11:36:17.375510 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Jan 24 11:36:17.375929 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Jan 24 11:36:17.376318 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Jan 24 11:36:17.376604 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Jan 24 11:36:17.377189 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Jan 24 11:36:17.377473 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Jan 24 11:36:17.377877 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Jan 24 11:36:17.378233 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Jan 24 11:36:17.378534 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Jan 24 11:36:17.378972 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Jan 24 11:36:17.379351 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Jan 24 11:36:17.379653 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Jan 24 11:36:17.380072 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Jan 24 11:36:17.380445 kernel: pci 0000:00:1f.0: quirk_ich7_lpc+0x0/0xc0 took 17578 usecs Jan 24 11:36:17.380954 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Jan 24 11:36:17.381331 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Jan 24 11:36:17.381623 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Jan 24 11:36:17.382060 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Jan 24 11:36:17.382434 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Jan 24 11:36:17.382453 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jan 24 11:36:17.382474 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jan 24 11:36:17.382487 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 24 11:36:17.382501 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jan 24 11:36:17.382512 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Jan 24 11:36:17.382523 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Jan 24 11:36:17.382535 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Jan 24 11:36:17.382551 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Jan 24 11:36:17.382567 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Jan 24 11:36:17.382578 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Jan 24 11:36:17.382592 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Jan 24 11:36:17.382606 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Jan 24 11:36:17.382617 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Jan 24 11:36:17.382628 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Jan 24 11:36:17.382642 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Jan 24 11:36:17.382659 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Jan 24 11:36:17.382673 kernel: iommu: Default domain type: Translated Jan 24 11:36:17.382684 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 24 11:36:17.382695 kernel: efivars: Registered efivars operations Jan 24 11:36:17.382707 kernel: PCI: Using ACPI for IRQ routing Jan 24 11:36:17.382843 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 24 11:36:17.382858 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Jan 24 11:36:17.382877 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Jan 24 11:36:17.382889 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Jan 24 11:36:17.382900 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Jan 24 11:36:17.382911 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Jan 24 11:36:17.382924 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Jan 24 11:36:17.382937 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Jan 24 11:36:17.382951 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Jan 24 11:36:17.383324 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Jan 24 11:36:17.383611 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Jan 24 11:36:17.384171 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 24 11:36:17.384194 kernel: vgaarb: loaded Jan 24 11:36:17.384206 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jan 24 11:36:17.384218 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jan 24 11:36:17.384230 kernel: clocksource: Switched to clocksource kvm-clock Jan 24 11:36:17.384248 kernel: VFS: Disk quotas dquot_6.6.0 Jan 24 11:36:17.384261 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 24 11:36:17.384274 kernel: pnp: PnP ACPI init Jan 24 11:36:17.384570 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Jan 24 11:36:17.384590 kernel: pnp: PnP ACPI: found 6 devices Jan 24 11:36:17.384604 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 24 11:36:17.384618 kernel: NET: Registered PF_INET protocol family Jan 24 11:36:17.384636 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 24 11:36:17.384649 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 24 11:36:17.384688 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 24 11:36:17.384705 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 24 11:36:17.384847 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 24 11:36:17.384862 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 24 11:36:17.384881 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 24 11:36:17.384894 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 24 11:36:17.384908 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 24 11:36:17.384921 kernel: NET: Registered PF_XDP protocol family Jan 24 11:36:17.385292 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Jan 24 11:36:17.385578 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Jan 24 11:36:17.386010 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jan 24 11:36:17.386364 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jan 24 11:36:17.386636 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jan 24 11:36:17.387093 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Jan 24 11:36:17.387450 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Jan 24 11:36:17.387861 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Jan 24 11:36:17.387883 kernel: PCI: CLS 0 bytes, default 64 Jan 24 11:36:17.387897 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Jan 24 11:36:17.387917 kernel: Initialise system trusted keyrings Jan 24 11:36:17.387930 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 24 11:36:17.387944 kernel: Key type asymmetric registered Jan 24 11:36:17.387956 kernel: Asymmetric key parser 'x509' registered Jan 24 11:36:17.387970 kernel: hrtimer: interrupt took 11788368 ns Jan 24 11:36:17.387984 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jan 24 11:36:17.387997 kernel: io scheduler mq-deadline registered Jan 24 11:36:17.388015 kernel: io scheduler kyber registered Jan 24 11:36:17.388029 kernel: io scheduler bfq registered Jan 24 11:36:17.388042 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 24 11:36:17.388058 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Jan 24 11:36:17.388079 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Jan 24 11:36:17.388093 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Jan 24 11:36:17.388187 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 24 11:36:17.388201 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 24 11:36:17.388219 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jan 24 11:36:17.388232 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 24 11:36:17.388246 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 24 11:36:17.388549 kernel: rtc_cmos 00:04: RTC can wake from S4 Jan 24 11:36:17.388570 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jan 24 11:36:17.389024 kernel: rtc_cmos 00:04: registered as rtc0 Jan 24 11:36:17.389377 kernel: rtc_cmos 00:04: setting system clock to 2026-01-24T11:36:01 UTC (1769254561) Jan 24 11:36:17.389648 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Jan 24 11:36:17.389668 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jan 24 11:36:17.389686 kernel: efifb: probing for efifb Jan 24 11:36:17.389700 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Jan 24 11:36:17.389713 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Jan 24 11:36:17.389863 kernel: efifb: scrolling: redraw Jan 24 11:36:17.389876 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Jan 24 11:36:17.389889 kernel: Console: switching to colour frame buffer device 160x50 Jan 24 11:36:17.389903 kernel: fb0: EFI VGA frame buffer device Jan 24 11:36:17.389916 kernel: pstore: Using crash dump compression: deflate Jan 24 11:36:17.389933 kernel: pstore: Registered efi_pstore as persistent store backend Jan 24 11:36:17.389945 kernel: NET: Registered PF_INET6 protocol family Jan 24 11:36:17.389957 kernel: Segment Routing with IPv6 Jan 24 11:36:17.389971 kernel: In-situ OAM (IOAM) with IPv6 Jan 24 11:36:17.389984 kernel: NET: Registered PF_PACKET protocol family Jan 24 11:36:17.389999 kernel: Key type dns_resolver registered Jan 24 11:36:17.390011 kernel: IPI shorthand broadcast: enabled Jan 24 11:36:17.390027 kernel: sched_clock: Marking stable (15758141530, 4089690055)->(22191793524, -2343961939) Jan 24 11:36:17.390040 kernel: registered taskstats version 1 Jan 24 11:36:17.390054 kernel: Loading compiled-in X.509 certificates Jan 24 11:36:17.390068 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.66-flatcar: a97c6138cc1b5c46f82656a7e055bcfc44b38b5c' Jan 24 11:36:17.390082 kernel: Demotion targets for Node 0: null Jan 24 11:36:17.390093 kernel: Key type .fscrypt registered Jan 24 11:36:17.390188 kernel: Key type fscrypt-provisioning registered Jan 24 11:36:17.390209 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 24 11:36:17.390222 kernel: ima: Allocated hash algorithm: sha1 Jan 24 11:36:17.390234 kernel: ima: No architecture policies found Jan 24 11:36:17.390245 kernel: clk: Disabling unused clocks Jan 24 11:36:17.390258 kernel: Freeing unused kernel image (initmem) memory: 15536K Jan 24 11:36:17.390272 kernel: Write protecting the kernel read-only data: 47104k Jan 24 11:36:17.390286 kernel: Freeing unused kernel image (rodata/data gap) memory: 1124K Jan 24 11:36:17.390303 kernel: Run /init as init process Jan 24 11:36:17.390316 kernel: with arguments: Jan 24 11:36:17.390328 kernel: /init Jan 24 11:36:17.390340 kernel: with environment: Jan 24 11:36:17.390353 kernel: HOME=/ Jan 24 11:36:17.390365 kernel: TERM=linux Jan 24 11:36:17.390377 kernel: SCSI subsystem initialized Jan 24 11:36:17.390392 kernel: libata version 3.00 loaded. Jan 24 11:36:17.390694 kernel: ahci 0000:00:1f.2: version 3.0 Jan 24 11:36:17.390842 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Jan 24 11:36:17.391217 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Jan 24 11:36:17.391527 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Jan 24 11:36:17.391979 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Jan 24 11:36:17.392651 kernel: scsi host0: ahci Jan 24 11:36:17.393185 kernel: scsi host1: ahci Jan 24 11:36:17.393500 kernel: scsi host2: ahci Jan 24 11:36:17.393940 kernel: scsi host3: ahci Jan 24 11:36:17.394339 kernel: scsi host4: ahci Jan 24 11:36:17.394648 kernel: scsi host5: ahci Jan 24 11:36:17.394676 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 26 lpm-pol 1 Jan 24 11:36:17.394690 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 26 lpm-pol 1 Jan 24 11:36:17.394703 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 26 lpm-pol 1 Jan 24 11:36:17.394848 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 26 lpm-pol 1 Jan 24 11:36:17.394864 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 26 lpm-pol 1 Jan 24 11:36:17.394879 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 26 lpm-pol 1 Jan 24 11:36:17.394892 kernel: ata4: SATA link down (SStatus 0 SControl 300) Jan 24 11:36:17.394911 kernel: ata5: SATA link down (SStatus 0 SControl 300) Jan 24 11:36:17.394925 kernel: ata6: SATA link down (SStatus 0 SControl 300) Jan 24 11:36:17.394938 kernel: ata2: SATA link down (SStatus 0 SControl 300) Jan 24 11:36:17.394955 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Jan 24 11:36:17.394969 kernel: ata1: SATA link down (SStatus 0 SControl 300) Jan 24 11:36:17.394983 kernel: ata3.00: LPM support broken, forcing max_power Jan 24 11:36:17.394996 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Jan 24 11:36:17.395012 kernel: ata3.00: applying bridge limits Jan 24 11:36:17.395024 kernel: ata3.00: LPM support broken, forcing max_power Jan 24 11:36:17.395036 kernel: ata3.00: configured for UDMA/100 Jan 24 11:36:17.395512 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Jan 24 11:36:17.395968 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Jan 24 11:36:17.396392 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Jan 24 11:36:17.396421 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 24 11:36:17.396711 kernel: virtio_blk virtio1: [vda] 27000832 512-byte logical blocks (13.8 GB/12.9 GiB) Jan 24 11:36:17.397268 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jan 24 11:36:17.397291 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 24 11:36:17.397305 kernel: GPT:16515071 != 27000831 Jan 24 11:36:17.397318 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 24 11:36:17.397337 kernel: GPT:16515071 != 27000831 Jan 24 11:36:17.397350 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 24 11:36:17.397363 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 24 11:36:17.397377 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 24 11:36:17.397390 kernel: device-mapper: uevent: version 1.0.3 Jan 24 11:36:17.397405 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Jan 24 11:36:17.397418 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Jan 24 11:36:17.397434 kernel: raid6: avx2x4 gen() 12862 MB/s Jan 24 11:36:17.397445 kernel: raid6: avx2x2 gen() 11427 MB/s Jan 24 11:36:17.397458 kernel: raid6: avx2x1 gen() 15060 MB/s Jan 24 11:36:17.397471 kernel: raid6: using algorithm avx2x1 gen() 15060 MB/s Jan 24 11:36:17.397486 kernel: raid6: .... xor() 10206 MB/s, rmw enabled Jan 24 11:36:17.397499 kernel: raid6: using avx2x2 recovery algorithm Jan 24 11:36:17.397513 kernel: xor: automatically using best checksumming function avx Jan 24 11:36:17.397526 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 24 11:36:17.397544 kernel: BTRFS: device fsid d3bd77fc-0f38-45e2-bb37-1f1b4d0917b8 devid 1 transid 34 /dev/mapper/usr (253:0) scanned by mount (181) Jan 24 11:36:17.397559 kernel: BTRFS info (device dm-0): first mount of filesystem d3bd77fc-0f38-45e2-bb37-1f1b4d0917b8 Jan 24 11:36:17.397572 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 24 11:36:17.397586 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 24 11:36:17.397599 kernel: BTRFS info (device dm-0): enabling free space tree Jan 24 11:36:17.397612 kernel: loop: module loaded Jan 24 11:36:17.397625 kernel: loop0: detected capacity change from 0 to 100552 Jan 24 11:36:17.397643 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 24 11:36:17.397658 systemd[1]: Successfully made /usr/ read-only. Jan 24 11:36:17.397675 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 24 11:36:17.397689 systemd[1]: Detected virtualization kvm. Jan 24 11:36:17.397703 systemd[1]: Detected architecture x86-64. Jan 24 11:36:17.397846 systemd[1]: Running in initrd. Jan 24 11:36:17.397862 systemd[1]: No hostname configured, using default hostname. Jan 24 11:36:17.397877 systemd[1]: Hostname set to . Jan 24 11:36:17.397891 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Jan 24 11:36:17.397904 systemd[1]: Queued start job for default target initrd.target. Jan 24 11:36:17.397913 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 24 11:36:17.397921 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 24 11:36:17.397935 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 24 11:36:17.397945 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 24 11:36:17.397953 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 24 11:36:17.397963 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 24 11:36:17.397972 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 24 11:36:17.397984 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 24 11:36:17.397992 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 24 11:36:17.398001 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Jan 24 11:36:17.398010 systemd[1]: Reached target paths.target - Path Units. Jan 24 11:36:17.398021 systemd[1]: Reached target slices.target - Slice Units. Jan 24 11:36:17.398030 systemd[1]: Reached target swap.target - Swaps. Jan 24 11:36:17.398038 systemd[1]: Reached target timers.target - Timer Units. Jan 24 11:36:17.398049 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 24 11:36:17.398058 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 24 11:36:17.398067 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 24 11:36:17.398076 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 24 11:36:17.398084 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Jan 24 11:36:17.398093 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 24 11:36:17.398165 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 24 11:36:17.398178 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 24 11:36:17.398187 systemd[1]: Reached target sockets.target - Socket Units. Jan 24 11:36:17.398196 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 24 11:36:17.398204 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 24 11:36:17.398213 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 24 11:36:17.398222 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 24 11:36:17.398231 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Jan 24 11:36:17.398242 systemd[1]: Starting systemd-fsck-usr.service... Jan 24 11:36:17.398251 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 24 11:36:17.398259 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 24 11:36:17.398269 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 24 11:36:17.398280 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 24 11:36:17.398289 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 24 11:36:17.398298 systemd[1]: Finished systemd-fsck-usr.service. Jan 24 11:36:17.398307 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 24 11:36:17.398446 systemd-journald[323]: Collecting audit messages is enabled. Jan 24 11:36:17.398473 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 24 11:36:17.398483 systemd-journald[323]: Journal started Jan 24 11:36:17.398500 systemd-journald[323]: Runtime Journal (/run/log/journal/2cde8e0cdecc42f0a016bfc5aa3311a3) is 6M, max 48M, 42M free. Jan 24 11:36:17.406867 systemd[1]: Started systemd-journald.service - Journal Service. Jan 24 11:36:17.416000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.443864 kernel: audit: type=1130 audit(1769254577.416:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.452072 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 24 11:36:17.480687 kernel: Bridge firewalling registered Jan 24 11:36:17.501306 kernel: audit: type=1130 audit(1769254577.487:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.487000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.464956 systemd-modules-load[324]: Inserted module 'br_netfilter' Jan 24 11:36:17.468231 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 24 11:36:17.491480 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 24 11:36:17.564993 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:36:17.601250 kernel: audit: type=1130 audit(1769254577.567:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.567000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.602710 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 24 11:36:17.638000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.654087 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 24 11:36:17.679240 kernel: audit: type=1130 audit(1769254577.638:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.680916 systemd-tmpfiles[335]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Jan 24 11:36:17.702420 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 24 11:36:17.753064 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 24 11:36:17.752000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.775014 kernel: audit: type=1130 audit(1769254577.752:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.792579 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 24 11:36:17.798000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.806005 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 24 11:36:17.841196 kernel: audit: type=1130 audit(1769254577.798:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.841239 kernel: audit: type=1334 audit(1769254577.803:8): prog-id=6 op=LOAD Jan 24 11:36:17.803000 audit: BPF prog-id=6 op=LOAD Jan 24 11:36:17.852703 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 24 11:36:17.878000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.900182 kernel: audit: type=1130 audit(1769254577.878:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.909695 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 24 11:36:17.964215 kernel: audit: type=1130 audit(1769254577.923:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.923000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:17.964351 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 24 11:36:18.222311 dracut-cmdline[362]: dracut-109 Jan 24 11:36:18.261504 dracut-cmdline[362]: Using kernel command line parameters: SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=7953d3c7acaad6ee79638a10c67ea9f0b3a8597919989b6fbf2f9a1742d4ba63 Jan 24 11:36:18.322697 systemd-resolved[357]: Positive Trust Anchors: Jan 24 11:36:18.322914 systemd-resolved[357]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 24 11:36:18.322922 systemd-resolved[357]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 24 11:36:18.322970 systemd-resolved[357]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 24 11:36:18.644430 systemd-resolved[357]: Defaulting to hostname 'linux'. Jan 24 11:36:18.658432 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 24 11:36:18.693249 kernel: audit: type=1130 audit(1769254578.657:11): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:18.657000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:18.685347 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 24 11:36:19.327055 kernel: Loading iSCSI transport class v2.0-870. Jan 24 11:36:19.420001 kernel: iscsi: registered transport (tcp) Jan 24 11:36:19.508400 kernel: iscsi: registered transport (qla4xxx) Jan 24 11:36:19.508974 kernel: QLogic iSCSI HBA Driver Jan 24 11:36:19.749698 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 24 11:36:19.915934 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 24 11:36:19.920000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:19.924486 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 24 11:36:21.477592 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 24 11:36:21.512000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:21.523674 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 24 11:36:21.584269 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 24 11:36:21.968288 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 24 11:36:22.010000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:22.071000 audit: BPF prog-id=7 op=LOAD Jan 24 11:36:22.073000 audit: BPF prog-id=8 op=LOAD Jan 24 11:36:22.076013 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 24 11:36:22.514243 systemd-udevd[588]: Using default interface naming scheme 'v257'. Jan 24 11:36:22.649568 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 24 11:36:22.675000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:22.683662 kernel: kauditd_printk_skb: 5 callbacks suppressed Jan 24 11:36:22.684648 kernel: audit: type=1130 audit(1769254582.675:17): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:22.696600 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 24 11:36:22.883309 dracut-pre-trigger[633]: rd.md=0: removing MD RAID activation Jan 24 11:36:23.100026 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 24 11:36:23.158516 kernel: audit: type=1130 audit(1769254583.105:18): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:23.158651 kernel: audit: type=1334 audit(1769254583.125:19): prog-id=9 op=LOAD Jan 24 11:36:23.105000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:23.125000 audit: BPF prog-id=9 op=LOAD Jan 24 11:36:23.143443 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 24 11:36:23.188305 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 24 11:36:23.231025 kernel: audit: type=1130 audit(1769254583.196:20): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:23.196000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:23.200892 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 24 11:36:23.386947 systemd-networkd[723]: lo: Link UP Jan 24 11:36:23.387013 systemd-networkd[723]: lo: Gained carrier Jan 24 11:36:23.390461 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 24 11:36:23.428437 kernel: audit: type=1130 audit(1769254583.405:21): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:23.405000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:23.406561 systemd[1]: Reached target network.target - Network. Jan 24 11:36:23.515277 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 24 11:36:23.563674 kernel: audit: type=1130 audit(1769254583.520:22): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:23.520000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:23.564097 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 24 11:36:23.736638 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jan 24 11:36:23.787382 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jan 24 11:36:23.866659 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jan 24 11:36:23.910264 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 24 11:36:23.926466 kernel: cryptd: max_cpu_qlen set to 1000 Jan 24 11:36:23.955499 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 24 11:36:23.968562 systemd-networkd[723]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 24 11:36:23.968571 systemd-networkd[723]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 24 11:36:23.974302 systemd-networkd[723]: eth0: Link UP Jan 24 11:36:24.075910 kernel: audit: type=1131 audit(1769254584.034:23): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:24.034000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:23.975546 systemd-networkd[723]: eth0: Gained carrier Jan 24 11:36:23.975567 systemd-networkd[723]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 24 11:36:24.023297 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 24 11:36:24.023485 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:36:24.074324 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 24 11:36:24.158417 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 24 11:36:24.247586 disk-uuid[771]: Primary Header is updated. Jan 24 11:36:24.247586 disk-uuid[771]: Secondary Entries is updated. Jan 24 11:36:24.247586 disk-uuid[771]: Secondary Header is updated. Jan 24 11:36:24.201218 systemd-networkd[723]: eth0: DHCPv4 address 10.0.0.26/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 24 11:36:24.347940 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Jan 24 11:36:24.371281 kernel: AES CTR mode by8 optimization enabled Jan 24 11:36:24.515520 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:36:24.581347 kernel: audit: type=1130 audit(1769254584.547:24): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:24.547000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:25.072442 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 24 11:36:25.111000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:25.116059 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 24 11:36:25.177264 kernel: audit: type=1130 audit(1769254585.111:25): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:25.198395 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 24 11:36:25.217868 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 24 11:36:25.364311 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 24 11:36:25.512993 disk-uuid[773]: Warning: The kernel is still using the old partition table. Jan 24 11:36:25.512993 disk-uuid[773]: The new table will be used at the next reboot or after you Jan 24 11:36:25.512993 disk-uuid[773]: run partprobe(8) or kpartx(8) Jan 24 11:36:25.512993 disk-uuid[773]: The operation has completed successfully. Jan 24 11:36:25.689260 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 24 11:36:25.690132 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 24 11:36:25.787369 kernel: audit: type=1130 audit(1769254585.724:26): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:25.724000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:25.725000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:25.732990 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 24 11:36:25.811627 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 24 11:36:25.865000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:25.865593 systemd-networkd[723]: eth0: Gained IPv6LL Jan 24 11:36:26.084952 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (861) Jan 24 11:36:26.108223 kernel: BTRFS info (device vda6): first mount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:36:26.108508 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 24 11:36:26.188356 kernel: BTRFS info (device vda6): turning on async discard Jan 24 11:36:26.188602 kernel: BTRFS info (device vda6): enabling free space tree Jan 24 11:36:26.301538 kernel: BTRFS info (device vda6): last unmount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:36:26.325419 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 24 11:36:26.343000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:26.359387 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 24 11:36:27.798101 ignition[880]: Ignition 2.24.0 Jan 24 11:36:27.798596 ignition[880]: Stage: fetch-offline Jan 24 11:36:27.798930 ignition[880]: no configs at "/usr/lib/ignition/base.d" Jan 24 11:36:27.798949 ignition[880]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:36:27.800030 ignition[880]: parsed url from cmdline: "" Jan 24 11:36:27.800036 ignition[880]: no config URL provided Jan 24 11:36:27.801509 ignition[880]: reading system config file "/usr/lib/ignition/user.ign" Jan 24 11:36:27.801536 ignition[880]: no config at "/usr/lib/ignition/user.ign" Jan 24 11:36:27.801606 ignition[880]: op(1): [started] loading QEMU firmware config module Jan 24 11:36:27.801615 ignition[880]: op(1): executing: "modprobe" "qemu_fw_cfg" Jan 24 11:36:28.047640 ignition[880]: op(1): [finished] loading QEMU firmware config module Jan 24 11:36:29.425989 ignition[880]: parsing config with SHA512: 28ed4f20d8f8808c751bb43a7cf6257a5fe49ee1d1c5ba1211336b26471f0e4ac2244f2d36f11735c7aedf0e09c8ae49d8ba23db50457ff663f3ab2eb1393ddd Jan 24 11:36:29.667263 unknown[880]: fetched base config from "system" Jan 24 11:36:29.667358 unknown[880]: fetched user config from "qemu" Jan 24 11:36:29.684631 ignition[880]: fetch-offline: fetch-offline passed Jan 24 11:36:29.684993 ignition[880]: Ignition finished successfully Jan 24 11:36:29.711657 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 24 11:36:29.780285 kernel: kauditd_printk_skb: 3 callbacks suppressed Jan 24 11:36:29.780490 kernel: audit: type=1130 audit(1769254589.718:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:29.718000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:29.722670 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jan 24 11:36:29.801029 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 24 11:36:30.199297 ignition[891]: Ignition 2.24.0 Jan 24 11:36:30.199362 ignition[891]: Stage: kargs Jan 24 11:36:30.199872 ignition[891]: no configs at "/usr/lib/ignition/base.d" Jan 24 11:36:30.199886 ignition[891]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:36:30.231057 ignition[891]: kargs: kargs passed Jan 24 11:36:30.231323 ignition[891]: Ignition finished successfully Jan 24 11:36:30.252307 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 24 11:36:30.272000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:30.282894 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 24 11:36:30.307692 kernel: audit: type=1130 audit(1769254590.272:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:30.484315 ignition[899]: Ignition 2.24.0 Jan 24 11:36:30.484413 ignition[899]: Stage: disks Jan 24 11:36:30.485964 ignition[899]: no configs at "/usr/lib/ignition/base.d" Jan 24 11:36:30.486072 ignition[899]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:36:30.491268 ignition[899]: disks: disks passed Jan 24 11:36:30.491349 ignition[899]: Ignition finished successfully Jan 24 11:36:30.534428 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 24 11:36:30.578912 kernel: audit: type=1130 audit(1769254590.546:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:30.546000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:30.549003 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 24 11:36:30.591545 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 24 11:36:30.640447 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 24 11:36:30.654478 systemd[1]: Reached target sysinit.target - System Initialization. Jan 24 11:36:30.708611 systemd[1]: Reached target basic.target - Basic System. Jan 24 11:36:30.750508 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 24 11:36:30.903891 systemd-fsck[909]: ROOT: clean, 15/456736 files, 38230/456704 blocks Jan 24 11:36:30.929312 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 24 11:36:30.994894 kernel: audit: type=1130 audit(1769254590.957:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:30.957000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:30.961917 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 24 11:36:31.647602 kernel: EXT4-fs (vda9): mounted filesystem 04920273-eebf-4ad5-828c-7340043c8075 r/w with ordered data mode. Quota mode: none. Jan 24 11:36:31.653110 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 24 11:36:31.682420 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 24 11:36:31.719578 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 24 11:36:31.751624 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 24 11:36:31.762493 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 24 11:36:31.788001 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (918) Jan 24 11:36:31.773941 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 24 11:36:31.773997 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 24 11:36:31.807051 kernel: BTRFS info (device vda6): first mount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:36:31.807285 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 24 11:36:31.876589 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 24 11:36:31.894293 kernel: BTRFS info (device vda6): turning on async discard Jan 24 11:36:31.894335 kernel: BTRFS info (device vda6): enabling free space tree Jan 24 11:36:31.907588 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 24 11:36:31.912430 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 24 11:36:32.717697 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 24 11:36:32.800656 kernel: audit: type=1130 audit(1769254592.717:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:32.717000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:32.751519 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 24 11:36:32.829468 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 24 11:36:32.863049 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 24 11:36:32.896469 kernel: BTRFS info (device vda6): last unmount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:36:33.086696 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 24 11:36:33.112000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:33.154029 kernel: audit: type=1130 audit(1769254593.112:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:33.260639 ignition[1015]: INFO : Ignition 2.24.0 Jan 24 11:36:33.260639 ignition[1015]: INFO : Stage: mount Jan 24 11:36:33.260639 ignition[1015]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 24 11:36:33.260639 ignition[1015]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:36:33.300575 ignition[1015]: INFO : mount: mount passed Jan 24 11:36:33.306713 ignition[1015]: INFO : Ignition finished successfully Jan 24 11:36:33.321091 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 24 11:36:33.381297 kernel: audit: type=1130 audit(1769254593.327:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:33.327000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:33.331906 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 24 11:36:33.443960 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 24 11:36:33.606518 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1028) Jan 24 11:36:33.633052 kernel: BTRFS info (device vda6): first mount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:36:33.633152 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 24 11:36:33.684550 kernel: BTRFS info (device vda6): turning on async discard Jan 24 11:36:33.684652 kernel: BTRFS info (device vda6): enabling free space tree Jan 24 11:36:33.690185 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 24 11:36:34.166558 ignition[1045]: INFO : Ignition 2.24.0 Jan 24 11:36:34.166558 ignition[1045]: INFO : Stage: files Jan 24 11:36:34.166558 ignition[1045]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 24 11:36:34.166558 ignition[1045]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:36:34.205952 ignition[1045]: DEBUG : files: compiled without relabeling support, skipping Jan 24 11:36:34.205952 ignition[1045]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 24 11:36:34.205952 ignition[1045]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 24 11:36:34.205952 ignition[1045]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 24 11:36:34.205952 ignition[1045]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 24 11:36:34.205952 ignition[1045]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 24 11:36:34.205099 unknown[1045]: wrote ssh authorized keys file for user: core Jan 24 11:36:34.361541 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 24 11:36:34.361541 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Jan 24 11:36:34.669586 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 24 11:36:35.438475 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 24 11:36:35.438475 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 24 11:36:35.438475 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 24 11:36:35.438475 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Jan 24 11:36:35.504985 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.34.1-x86-64.raw: attempt #1 Jan 24 11:36:35.933473 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 24 11:36:40.829652 ignition[1045]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Jan 24 11:36:40.829652 ignition[1045]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 24 11:36:40.860934 ignition[1045]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 24 11:36:40.860934 ignition[1045]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 24 11:36:40.860934 ignition[1045]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 24 11:36:40.860934 ignition[1045]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Jan 24 11:36:40.860934 ignition[1045]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 24 11:36:40.860934 ignition[1045]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 24 11:36:40.860934 ignition[1045]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Jan 24 11:36:40.860934 ignition[1045]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Jan 24 11:36:41.122860 ignition[1045]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Jan 24 11:36:41.186112 ignition[1045]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jan 24 11:36:41.202433 ignition[1045]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Jan 24 11:36:41.202433 ignition[1045]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Jan 24 11:36:41.202433 ignition[1045]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Jan 24 11:36:41.202433 ignition[1045]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 24 11:36:41.202433 ignition[1045]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 24 11:36:41.202433 ignition[1045]: INFO : files: files passed Jan 24 11:36:41.202433 ignition[1045]: INFO : Ignition finished successfully Jan 24 11:36:41.291980 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 24 11:36:41.350066 kernel: audit: type=1130 audit(1769254601.305:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.305000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.351971 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 24 11:36:41.366915 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 24 11:36:41.435179 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 24 11:36:41.442124 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 24 11:36:41.458436 initrd-setup-root-after-ignition[1076]: grep: /sysroot/oem/oem-release: No such file or directory Jan 24 11:36:41.521183 kernel: audit: type=1130 audit(1769254601.457:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.521473 kernel: audit: type=1131 audit(1769254601.457:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.457000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.457000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.523888 initrd-setup-root-after-ignition[1082]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 24 11:36:41.537420 initrd-setup-root-after-ignition[1078]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 24 11:36:41.537420 initrd-setup-root-after-ignition[1078]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 24 11:36:41.623680 kernel: audit: type=1130 audit(1769254601.546:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.546000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.532166 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 24 11:36:41.609967 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 24 11:36:41.638995 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 24 11:36:41.803608 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 24 11:36:41.877626 kernel: audit: type=1130 audit(1769254601.817:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.877673 kernel: audit: type=1131 audit(1769254601.817:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.817000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.817000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:41.804310 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 24 11:36:41.818669 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 24 11:36:41.887645 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 24 11:36:41.942570 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 24 11:36:41.986084 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 24 11:36:42.347666 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 24 11:36:42.398401 kernel: audit: type=1130 audit(1769254602.358:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:42.358000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:42.363653 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 24 11:36:42.492683 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 24 11:36:42.493426 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 24 11:36:42.506423 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 24 11:36:42.558659 systemd[1]: Stopped target timers.target - Timer Units. Jan 24 11:36:42.659694 kernel: audit: type=1131 audit(1769254602.611:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:42.611000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:42.587016 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 24 11:36:42.590560 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 24 11:36:42.661334 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 24 11:36:42.680548 systemd[1]: Stopped target basic.target - Basic System. Jan 24 11:36:42.702688 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 24 11:36:42.722625 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 24 11:36:42.753401 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 24 11:36:42.770413 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Jan 24 11:36:42.793355 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 24 11:36:42.825524 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 24 11:36:42.869320 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 24 11:36:42.891074 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 24 11:36:42.913037 systemd[1]: Stopped target swap.target - Swaps. Jan 24 11:36:42.942541 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 24 11:36:42.985020 kernel: audit: type=1131 audit(1769254602.953:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:42.953000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:42.947071 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 24 11:36:42.981691 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 24 11:36:42.995860 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 24 11:36:43.011970 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 24 11:36:43.083156 kernel: audit: type=1131 audit(1769254603.057:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.057000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.015443 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 24 11:36:43.032561 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 24 11:36:43.035545 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 24 11:36:43.118000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.094531 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 24 11:36:43.095083 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 24 11:36:43.119449 systemd[1]: Stopped target paths.target - Path Units. Jan 24 11:36:43.144622 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 24 11:36:43.154902 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 24 11:36:43.171501 systemd[1]: Stopped target slices.target - Slice Units. Jan 24 11:36:43.204018 systemd[1]: Stopped target sockets.target - Socket Units. Jan 24 11:36:43.208984 systemd[1]: iscsid.socket: Deactivated successfully. Jan 24 11:36:43.297000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.209143 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 24 11:36:43.312000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.231463 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 24 11:36:43.231619 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 24 11:36:43.341000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.250505 systemd[1]: systemd-journald-audit.socket: Deactivated successfully. Jan 24 11:36:43.250639 systemd[1]: Closed systemd-journald-audit.socket - Journal Audit Socket. Jan 24 11:36:43.277395 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 24 11:36:43.277637 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 24 11:36:43.382000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.298955 systemd[1]: ignition-files.service: Deactivated successfully. Jan 24 11:36:43.396000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.299160 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 24 11:36:43.414000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.317900 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 24 11:36:43.323941 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 24 11:36:43.324142 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 24 11:36:43.363344 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 24 11:36:43.367910 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 24 11:36:43.368137 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 24 11:36:43.384341 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 24 11:36:43.509000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.509000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.384533 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 24 11:36:43.401214 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 24 11:36:43.401445 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 24 11:36:43.477537 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 24 11:36:43.491700 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 24 11:36:43.582139 ignition[1103]: INFO : Ignition 2.24.0 Jan 24 11:36:43.582139 ignition[1103]: INFO : Stage: umount Jan 24 11:36:43.596006 ignition[1103]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 24 11:36:43.596006 ignition[1103]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:36:43.596006 ignition[1103]: INFO : umount: umount passed Jan 24 11:36:43.596006 ignition[1103]: INFO : Ignition finished successfully Jan 24 11:36:43.630183 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 24 11:36:43.631491 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 24 11:36:43.645000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.631874 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 24 11:36:43.658000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.646465 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 24 11:36:43.646679 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 24 11:36:43.669104 systemd[1]: Stopped target network.target - Network. Jan 24 11:36:43.692000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.680922 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 24 11:36:43.703000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.681141 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 24 11:36:43.717000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.694147 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 24 11:36:43.731000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.694322 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 24 11:36:43.745000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.704671 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 24 11:36:43.704925 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 24 11:36:43.718898 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 24 11:36:43.718992 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 24 11:36:43.732051 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 24 11:36:43.732149 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 24 11:36:43.820000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.746668 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 24 11:36:43.759896 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 24 11:36:43.809025 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 24 11:36:43.853000 audit: BPF prog-id=6 op=UNLOAD Jan 24 11:36:43.809940 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 24 11:36:43.854452 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 24 11:36:43.884000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.863997 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 24 11:36:43.896400 systemd[1]: Stopped target network-pre.target - Preparation for Network. Jan 24 11:36:43.909514 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 24 11:36:43.909622 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 24 11:36:43.936167 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 24 11:36:43.966000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.946564 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 24 11:36:43.980000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.946670 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 24 11:36:44.007000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:43.967457 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 24 11:36:43.967590 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 24 11:36:43.981510 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 24 11:36:43.981621 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 24 11:36:44.013881 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 24 11:36:44.089000 audit: BPF prog-id=9 op=UNLOAD Jan 24 11:36:44.104452 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 24 11:36:44.104902 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 24 11:36:44.118000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.119996 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 24 11:36:44.120114 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 24 11:36:44.143549 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 24 11:36:44.143926 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 24 11:36:44.188190 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 24 11:36:44.201000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.189938 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 24 11:36:44.220000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.209372 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 24 11:36:44.236000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.209689 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 24 11:36:44.222966 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 24 11:36:44.321000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.223088 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 24 11:36:44.376000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.262971 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 24 11:36:44.423000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.282681 systemd[1]: systemd-network-generator.service: Deactivated successfully. Jan 24 11:36:44.282932 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Jan 24 11:36:44.323516 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 24 11:36:44.551000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.551000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.324612 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 24 11:36:44.403852 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 24 11:36:44.404437 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:36:44.481706 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 24 11:36:44.482221 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 24 11:36:44.819923 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 24 11:36:44.872000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:36:44.820396 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 24 11:36:44.878007 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 24 11:36:44.929679 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 24 11:36:45.182929 systemd[1]: Switching root. Jan 24 11:36:45.719530 systemd-journald[323]: Received SIGTERM from PID 1 (systemd). Jan 24 11:36:45.719917 systemd-journald[323]: Journal stopped Jan 24 11:37:05.810555 kernel: SELinux: policy capability network_peer_controls=1 Jan 24 11:37:05.811199 kernel: SELinux: policy capability open_perms=1 Jan 24 11:37:05.811314 kernel: SELinux: policy capability extended_socket_class=1 Jan 24 11:37:05.811470 kernel: SELinux: policy capability always_check_network=0 Jan 24 11:37:05.811491 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 24 11:37:05.811515 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 24 11:37:05.811539 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 24 11:37:05.811561 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 24 11:37:05.811579 kernel: SELinux: policy capability userspace_initial_context=0 Jan 24 11:37:05.811602 kernel: kauditd_printk_skb: 34 callbacks suppressed Jan 24 11:37:05.812183 kernel: audit: type=1403 audit(1769254608.593:81): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 24 11:37:05.812293 systemd[1]: Successfully loaded SELinux policy in 1.088307s. Jan 24 11:37:05.812968 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 58.017ms. Jan 24 11:37:05.813083 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 24 11:37:05.813177 systemd[1]: Detected virtualization kvm. Jan 24 11:37:05.813290 systemd[1]: Detected architecture x86-64. Jan 24 11:37:05.813548 systemd[1]: Detected first boot. Jan 24 11:37:05.813661 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Jan 24 11:37:05.813924 kernel: audit: type=1334 audit(1769254609.083:82): prog-id=10 op=LOAD Jan 24 11:37:05.813946 kernel: audit: type=1334 audit(1769254609.083:83): prog-id=10 op=UNLOAD Jan 24 11:37:05.813971 kernel: audit: type=1334 audit(1769254609.083:84): prog-id=11 op=LOAD Jan 24 11:37:05.814071 kernel: audit: type=1334 audit(1769254609.083:85): prog-id=11 op=UNLOAD Jan 24 11:37:05.814091 zram_generator::config[1148]: No configuration found. Jan 24 11:37:05.814111 kernel: Guest personality initialized and is inactive Jan 24 11:37:05.814449 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Jan 24 11:37:05.814549 kernel: Initialized host personality Jan 24 11:37:05.814571 kernel: NET: Registered PF_VSOCK protocol family Jan 24 11:37:05.814589 systemd[1]: Populated /etc with preset unit settings. Jan 24 11:37:05.814607 kernel: audit: type=1334 audit(1769254612.639:86): prog-id=12 op=LOAD Jan 24 11:37:05.814624 kernel: audit: type=1334 audit(1769254612.639:87): prog-id=3 op=UNLOAD Jan 24 11:37:05.814643 kernel: audit: type=1334 audit(1769254612.812:88): prog-id=13 op=LOAD Jan 24 11:37:05.815250 kernel: audit: type=1334 audit(1769254612.814:89): prog-id=14 op=LOAD Jan 24 11:37:05.815470 kernel: audit: type=1334 audit(1769254612.814:90): prog-id=4 op=UNLOAD Jan 24 11:37:05.815492 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 24 11:37:05.815693 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 24 11:37:05.815885 kernel: kauditd_printk_skb: 3 callbacks suppressed Jan 24 11:37:05.816006 kernel: audit: type=1130 audit(1769254613.808:94): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:05.816029 kernel: audit: type=1131 audit(1769254613.873:95): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:05.816130 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 24 11:37:05.816158 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 24 11:37:05.816177 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 24 11:37:05.816464 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 24 11:37:05.816486 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 24 11:37:05.816504 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 24 11:37:05.816627 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 24 11:37:05.816653 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 24 11:37:05.816671 systemd[1]: Created slice user.slice - User and Session Slice. Jan 24 11:37:05.816947 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 24 11:37:05.816969 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 24 11:37:05.816987 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 24 11:37:05.817006 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 24 11:37:05.817190 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 24 11:37:05.817213 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 24 11:37:05.817231 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 24 11:37:05.817511 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 24 11:37:05.817532 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 24 11:37:05.817553 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 24 11:37:05.817663 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 24 11:37:05.817685 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 24 11:37:05.817703 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 24 11:37:05.817902 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 24 11:37:05.818027 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 24 11:37:05.818048 systemd[1]: Reached target remote-veritysetup.target - Remote Verity Protected Volumes. Jan 24 11:37:05.818221 systemd[1]: Reached target slices.target - Slice Units. Jan 24 11:37:05.818241 systemd[1]: Reached target swap.target - Swaps. Jan 24 11:37:05.818259 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 24 11:37:05.818276 kernel: clocksource: Long readout interval, skipping watchdog check: cs_nsec: 1921163728 wd_nsec: 1921164131 Jan 24 11:37:05.818471 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 24 11:37:05.818585 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Jan 24 11:37:05.818608 kernel: audit: type=1335 audit(1769254621.014:96): pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Jan 24 11:37:05.818626 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 24 11:37:05.818893 systemd[1]: Listening on systemd-mountfsd.socket - DDI File System Mounter Socket. Jan 24 11:37:05.818914 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 24 11:37:05.818930 systemd[1]: Listening on systemd-nsresourced.socket - Namespace Resource Manager Socket. Jan 24 11:37:05.818947 systemd[1]: Listening on systemd-oomd.socket - Userspace Out-Of-Memory (OOM) Killer Socket. Jan 24 11:37:05.819060 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 24 11:37:05.819079 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 24 11:37:05.819256 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 24 11:37:05.819460 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 24 11:37:05.819482 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 24 11:37:05.819586 systemd[1]: Mounting media.mount - External Media Directory... Jan 24 11:37:05.819690 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:37:05.819881 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 24 11:37:05.819903 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 24 11:37:05.819923 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 24 11:37:05.820028 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 24 11:37:05.820049 systemd[1]: Reached target machines.target - Containers. Jan 24 11:37:05.820068 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 24 11:37:05.820173 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 24 11:37:05.820197 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 24 11:37:05.820215 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 24 11:37:05.820320 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 24 11:37:05.820510 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 24 11:37:05.820534 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 24 11:37:05.820637 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 24 11:37:05.820911 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 24 11:37:05.820935 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 24 11:37:05.821038 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 24 11:37:05.821058 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 24 11:37:05.821175 kernel: audit: type=1131 audit(1769254624.805:97): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:05.821198 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 24 11:37:05.821217 systemd[1]: Stopped systemd-fsck-usr.service. Jan 24 11:37:05.821335 kernel: audit: type=1131 audit(1769254624.920:98): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:05.821467 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 24 11:37:05.821490 kernel: audit: type=1334 audit(1769254625.022:99): prog-id=14 op=UNLOAD Jan 24 11:37:05.821680 kernel: audit: type=1334 audit(1769254625.022:100): prog-id=13 op=UNLOAD Jan 24 11:37:05.821700 kernel: audit: type=1334 audit(1769254625.068:101): prog-id=15 op=LOAD Jan 24 11:37:05.821961 kernel: audit: type=1334 audit(1769254625.104:102): prog-id=16 op=LOAD Jan 24 11:37:05.821984 kernel: audit: type=1334 audit(1769254625.105:103): prog-id=17 op=LOAD Jan 24 11:37:05.822001 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 24 11:37:05.822021 kernel: fuse: init (API version 7.41) Jan 24 11:37:05.822041 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 24 11:37:05.822156 kernel: ACPI: bus type drm_connector registered Jan 24 11:37:05.822177 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 24 11:37:05.822237 systemd-journald[1234]: Collecting audit messages is enabled. Jan 24 11:37:05.822530 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 24 11:37:05.822553 kernel: audit: type=1305 audit(1769254625.800:104): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Jan 24 11:37:05.822674 systemd-journald[1234]: Journal started Jan 24 11:37:05.822709 systemd-journald[1234]: Runtime Journal (/run/log/journal/2cde8e0cdecc42f0a016bfc5aa3311a3) is 6M, max 48M, 42M free. Jan 24 11:37:01.014000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Jan 24 11:37:04.805000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:04.920000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:05.022000 audit: BPF prog-id=14 op=UNLOAD Jan 24 11:37:05.022000 audit: BPF prog-id=13 op=UNLOAD Jan 24 11:37:05.068000 audit: BPF prog-id=15 op=LOAD Jan 24 11:37:05.104000 audit: BPF prog-id=16 op=LOAD Jan 24 11:37:05.105000 audit: BPF prog-id=17 op=LOAD Jan 24 11:37:05.800000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Jan 24 11:36:51.753247 systemd[1]: Queued start job for default target multi-user.target. Jan 24 11:36:52.870937 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jan 24 11:36:52.920629 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 24 11:36:52.924150 systemd[1]: systemd-journald.service: Consumed 4.840s CPU time. Jan 24 11:37:05.957182 kernel: audit: type=1300 audit(1769254625.800:104): arch=c000003e syscall=46 success=yes exit=60 a0=3 a1=7fff5b81cde0 a2=4000 a3=0 items=0 ppid=1 pid=1234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:37:05.800000 audit[1234]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=3 a1=7fff5b81cde0 a2=4000 a3=0 items=0 ppid=1 pid=1234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:37:05.800000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Jan 24 11:37:05.998240 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Jan 24 11:37:06.112947 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 24 11:37:06.139203 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:37:06.159475 systemd[1]: Started systemd-journald.service - Journal Service. Jan 24 11:37:06.173000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.177970 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 24 11:37:06.185623 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:37:06.185684 kernel: audit: type=1130 audit(1769254626.173:105): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.229228 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 24 11:37:06.246704 systemd[1]: Mounted media.mount - External Media Directory. Jan 24 11:37:06.286934 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 24 11:37:06.381083 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 24 11:37:06.458100 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 24 11:37:06.510515 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 24 11:37:06.538000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.601997 kernel: audit: type=1130 audit(1769254626.538:106): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.603534 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 24 11:37:06.670000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.674490 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 24 11:37:06.675146 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 24 11:37:06.722313 kernel: audit: type=1130 audit(1769254626.670:107): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.722000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.751466 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 24 11:37:06.752057 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 24 11:37:06.749000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.928060 kernel: audit: type=1130 audit(1769254626.722:108): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.928578 kernel: audit: type=1131 audit(1769254626.749:109): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.928626 kernel: audit: type=1130 audit(1769254626.921:110): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.921000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.932328 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 24 11:37:06.936246 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 24 11:37:07.064193 kernel: audit: type=1131 audit(1769254626.921:111): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:06.921000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.063000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.066955 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 24 11:37:07.068092 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 24 11:37:07.190575 kernel: audit: type=1130 audit(1769254627.063:112): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.192099 kernel: audit: type=1131 audit(1769254627.064:113): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.064000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.192000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.224281 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 24 11:37:07.226532 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 24 11:37:07.194000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.279257 kernel: audit: type=1130 audit(1769254627.192:114): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.267000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.267000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.278926 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 24 11:37:07.287179 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 24 11:37:07.301000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.301000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.304320 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 24 11:37:07.352000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.360580 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 24 11:37:07.385000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.403935 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 24 11:37:07.439000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.447319 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Jan 24 11:37:07.479000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-load-credentials comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.493682 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 24 11:37:07.523000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:07.773648 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 24 11:37:07.803590 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Jan 24 11:37:07.833688 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 24 11:37:07.874160 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 24 11:37:07.888243 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 24 11:37:07.888305 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 24 11:37:07.915691 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Jan 24 11:37:07.957293 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 24 11:37:07.958500 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 24 11:37:07.966649 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 24 11:37:07.984635 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 24 11:37:08.006998 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 24 11:37:08.042936 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 24 11:37:08.059690 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 24 11:37:08.067096 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 24 11:37:08.091711 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 24 11:37:08.120176 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 24 11:37:08.145919 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 24 11:37:08.150925 systemd-journald[1234]: Time spent on flushing to /var/log/journal/2cde8e0cdecc42f0a016bfc5aa3311a3 is 151.852ms for 1236 entries. Jan 24 11:37:08.150925 systemd-journald[1234]: System Journal (/var/log/journal/2cde8e0cdecc42f0a016bfc5aa3311a3) is 8M, max 163.5M, 155.5M free. Jan 24 11:37:08.453593 systemd-journald[1234]: Received client request to flush runtime journal. Jan 24 11:37:08.186556 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 24 11:37:08.526317 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 24 11:37:08.575000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:08.596043 kernel: loop1: detected capacity change from 0 to 50784 Jan 24 11:37:08.604631 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 24 11:37:08.627000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:08.635652 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 24 11:37:08.661000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:08.715617 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 24 11:37:08.787233 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Jan 24 11:37:09.195037 kernel: loop2: detected capacity change from 0 to 111560 Jan 24 11:37:09.257614 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 24 11:37:09.261230 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Jan 24 11:37:09.280000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:09.358549 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 24 11:37:09.376000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:09.383000 audit: BPF prog-id=18 op=LOAD Jan 24 11:37:09.384000 audit: BPF prog-id=19 op=LOAD Jan 24 11:37:09.384000 audit: BPF prog-id=20 op=LOAD Jan 24 11:37:09.386969 systemd[1]: Starting systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer... Jan 24 11:37:09.419000 audit: BPF prog-id=21 op=LOAD Jan 24 11:37:09.454195 kernel: loop3: detected capacity change from 0 to 219144 Jan 24 11:37:09.456634 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 24 11:37:09.499093 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 24 11:37:09.557000 audit: BPF prog-id=22 op=LOAD Jan 24 11:37:09.557000 audit: BPF prog-id=23 op=LOAD Jan 24 11:37:09.557000 audit: BPF prog-id=24 op=LOAD Jan 24 11:37:09.562189 systemd[1]: Starting systemd-nsresourced.service - Namespace Resource Manager... Jan 24 11:37:10.178000 audit: BPF prog-id=25 op=LOAD Jan 24 11:37:10.178000 audit: BPF prog-id=26 op=LOAD Jan 24 11:37:10.178000 audit: BPF prog-id=27 op=LOAD Jan 24 11:37:10.183608 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 24 11:37:10.218672 kernel: loop4: detected capacity change from 0 to 50784 Jan 24 11:37:10.301349 kernel: loop5: detected capacity change from 0 to 111560 Jan 24 11:37:10.446088 kernel: loop6: detected capacity change from 0 to 219144 Jan 24 11:37:10.558357 systemd-tmpfiles[1289]: ACLs are not supported, ignoring. Jan 24 11:37:10.562008 systemd-tmpfiles[1289]: ACLs are not supported, ignoring. Jan 24 11:37:10.757952 (sd-merge)[1293]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes.raw'. Jan 24 11:37:10.997095 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 24 11:37:11.009000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:11.076147 (sd-merge)[1293]: Merged extensions into '/usr'. Jan 24 11:37:11.104052 systemd[1]: Reload requested from client PID 1269 ('systemd-sysext') (unit systemd-sysext.service)... Jan 24 11:37:11.104079 systemd[1]: Reloading... Jan 24 11:37:11.127257 systemd-nsresourced[1291]: Not setting up BPF subsystem, as functionality has been disabled at compile time. Jan 24 11:37:11.604916 zram_generator::config[1358]: No configuration found. Jan 24 11:37:11.746891 systemd-resolved[1287]: Positive Trust Anchors: Jan 24 11:37:11.746918 systemd-resolved[1287]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 24 11:37:11.746925 systemd-resolved[1287]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 24 11:37:11.746973 systemd-resolved[1287]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 24 11:37:11.814924 systemd-resolved[1287]: Defaulting to hostname 'linux'. Jan 24 11:37:12.132274 systemd-oomd[1286]: No swap; memory pressure usage will be degraded Jan 24 11:37:12.889883 systemd[1]: Reloading finished in 1784 ms. Jan 24 11:37:13.320521 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 24 11:37:13.388516 kernel: kauditd_printk_skb: 26 callbacks suppressed Jan 24 11:37:13.403094 kernel: audit: type=1130 audit(1769254633.349:141): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.403277 kernel: audit: type=1130 audit(1769254633.401:142): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.349000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.401000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.351639 systemd[1]: Started systemd-nsresourced.service - Namespace Resource Manager. Jan 24 11:37:13.403949 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 24 11:37:13.453000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.455515 systemd[1]: Started systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer. Jan 24 11:37:13.506083 kernel: audit: type=1130 audit(1769254633.453:143): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.525000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.561104 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 24 11:37:13.596008 kernel: audit: type=1130 audit(1769254633.525:144): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.622000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.677496 kernel: audit: type=1130 audit(1769254633.622:145): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:13.711672 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 24 11:37:13.764708 systemd[1]: Starting ensure-sysext.service... Jan 24 11:37:13.777528 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 24 11:37:13.799000 audit: BPF prog-id=28 op=LOAD Jan 24 11:37:13.816893 kernel: audit: type=1334 audit(1769254633.799:146): prog-id=28 op=LOAD Jan 24 11:37:13.816947 kernel: audit: type=1334 audit(1769254633.799:147): prog-id=21 op=UNLOAD Jan 24 11:37:13.799000 audit: BPF prog-id=21 op=UNLOAD Jan 24 11:37:13.800000 audit: BPF prog-id=29 op=LOAD Jan 24 11:37:13.829193 systemd[1]: Reload requested from client PID 1373 ('systemctl') (unit ensure-sysext.service)... Jan 24 11:37:13.829211 systemd[1]: Reloading... Jan 24 11:37:13.844232 kernel: audit: type=1334 audit(1769254633.800:148): prog-id=29 op=LOAD Jan 24 11:37:13.845130 kernel: audit: type=1334 audit(1769254633.800:149): prog-id=22 op=UNLOAD Jan 24 11:37:13.845164 kernel: audit: type=1334 audit(1769254633.809:150): prog-id=30 op=LOAD Jan 24 11:37:13.800000 audit: BPF prog-id=22 op=UNLOAD Jan 24 11:37:13.809000 audit: BPF prog-id=30 op=LOAD Jan 24 11:37:13.809000 audit: BPF prog-id=31 op=LOAD Jan 24 11:37:13.809000 audit: BPF prog-id=23 op=UNLOAD Jan 24 11:37:13.809000 audit: BPF prog-id=24 op=UNLOAD Jan 24 11:37:13.810000 audit: BPF prog-id=32 op=LOAD Jan 24 11:37:13.810000 audit: BPF prog-id=25 op=UNLOAD Jan 24 11:37:13.811000 audit: BPF prog-id=33 op=LOAD Jan 24 11:37:13.811000 audit: BPF prog-id=34 op=LOAD Jan 24 11:37:13.811000 audit: BPF prog-id=26 op=UNLOAD Jan 24 11:37:13.811000 audit: BPF prog-id=27 op=UNLOAD Jan 24 11:37:13.813000 audit: BPF prog-id=35 op=LOAD Jan 24 11:37:13.813000 audit: BPF prog-id=15 op=UNLOAD Jan 24 11:37:13.813000 audit: BPF prog-id=36 op=LOAD Jan 24 11:37:13.813000 audit: BPF prog-id=37 op=LOAD Jan 24 11:37:13.814000 audit: BPF prog-id=16 op=UNLOAD Jan 24 11:37:13.814000 audit: BPF prog-id=17 op=UNLOAD Jan 24 11:37:13.815000 audit: BPF prog-id=38 op=LOAD Jan 24 11:37:13.815000 audit: BPF prog-id=18 op=UNLOAD Jan 24 11:37:13.815000 audit: BPF prog-id=39 op=LOAD Jan 24 11:37:13.815000 audit: BPF prog-id=40 op=LOAD Jan 24 11:37:13.816000 audit: BPF prog-id=19 op=UNLOAD Jan 24 11:37:13.816000 audit: BPF prog-id=20 op=UNLOAD Jan 24 11:37:14.446694 zram_generator::config[1408]: No configuration found. Jan 24 11:37:14.653188 systemd-tmpfiles[1374]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Jan 24 11:37:14.654596 systemd-tmpfiles[1374]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Jan 24 11:37:14.655972 systemd-tmpfiles[1374]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 24 11:37:14.763100 systemd-tmpfiles[1374]: ACLs are not supported, ignoring. Jan 24 11:37:14.763520 systemd-tmpfiles[1374]: ACLs are not supported, ignoring. Jan 24 11:37:14.890039 systemd-tmpfiles[1374]: Detected autofs mount point /boot during canonicalization of boot. Jan 24 11:37:14.907343 systemd-tmpfiles[1374]: Skipping /boot Jan 24 11:37:15.368373 systemd-tmpfiles[1374]: Detected autofs mount point /boot during canonicalization of boot. Jan 24 11:37:15.368564 systemd-tmpfiles[1374]: Skipping /boot Jan 24 11:37:16.088134 systemd[1]: Reloading finished in 2257 ms. Jan 24 11:37:16.321121 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 24 11:37:16.361000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:16.396000 audit: BPF prog-id=41 op=LOAD Jan 24 11:37:16.396000 audit: BPF prog-id=28 op=UNLOAD Jan 24 11:37:16.448000 audit: BPF prog-id=42 op=LOAD Jan 24 11:37:16.449000 audit: BPF prog-id=29 op=UNLOAD Jan 24 11:37:16.449000 audit: BPF prog-id=43 op=LOAD Jan 24 11:37:16.449000 audit: BPF prog-id=44 op=LOAD Jan 24 11:37:16.449000 audit: BPF prog-id=30 op=UNLOAD Jan 24 11:37:16.449000 audit: BPF prog-id=31 op=UNLOAD Jan 24 11:37:16.452000 audit: BPF prog-id=45 op=LOAD Jan 24 11:37:16.453000 audit: BPF prog-id=32 op=UNLOAD Jan 24 11:37:16.453000 audit: BPF prog-id=46 op=LOAD Jan 24 11:37:16.453000 audit: BPF prog-id=47 op=LOAD Jan 24 11:37:16.453000 audit: BPF prog-id=33 op=UNLOAD Jan 24 11:37:16.454000 audit: BPF prog-id=34 op=UNLOAD Jan 24 11:37:16.458000 audit: BPF prog-id=48 op=LOAD Jan 24 11:37:16.480000 audit: BPF prog-id=35 op=UNLOAD Jan 24 11:37:16.491000 audit: BPF prog-id=49 op=LOAD Jan 24 11:37:16.495000 audit: BPF prog-id=50 op=LOAD Jan 24 11:37:16.500000 audit: BPF prog-id=36 op=UNLOAD Jan 24 11:37:16.500000 audit: BPF prog-id=37 op=UNLOAD Jan 24 11:37:16.513000 audit: BPF prog-id=51 op=LOAD Jan 24 11:37:16.513000 audit: BPF prog-id=38 op=UNLOAD Jan 24 11:37:16.513000 audit: BPF prog-id=52 op=LOAD Jan 24 11:37:16.513000 audit: BPF prog-id=53 op=LOAD Jan 24 11:37:16.513000 audit: BPF prog-id=39 op=UNLOAD Jan 24 11:37:16.513000 audit: BPF prog-id=40 op=UNLOAD Jan 24 11:37:16.572068 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 24 11:37:16.587000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:16.677006 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 24 11:37:16.699289 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 24 11:37:16.759273 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 24 11:37:16.781137 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 24 11:37:16.806000 audit: BPF prog-id=8 op=UNLOAD Jan 24 11:37:16.806000 audit: BPF prog-id=7 op=UNLOAD Jan 24 11:37:16.808000 audit: BPF prog-id=54 op=LOAD Jan 24 11:37:16.809000 audit: BPF prog-id=55 op=LOAD Jan 24 11:37:16.831147 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 24 11:37:16.883279 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 24 11:37:16.950339 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:37:16.950925 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 24 11:37:16.966110 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 24 11:37:17.004032 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 24 11:37:17.040556 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 24 11:37:17.056000 audit[1455]: SYSTEM_BOOT pid=1455 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.054213 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 24 11:37:17.055139 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 24 11:37:17.055312 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 24 11:37:17.055696 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:37:17.084613 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:37:17.085197 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 24 11:37:17.085603 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 24 11:37:17.088641 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 24 11:37:17.089373 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 24 11:37:17.089632 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:37:17.096091 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 24 11:37:17.101000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.117588 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:37:17.119237 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 24 11:37:17.131363 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 24 11:37:17.159666 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 24 11:37:17.160391 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 24 11:37:17.160696 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 24 11:37:17.161196 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:37:17.168346 systemd-udevd[1454]: Using default interface naming scheme 'v257'. Jan 24 11:37:17.178596 systemd[1]: Finished ensure-sysext.service. Jan 24 11:37:17.191000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.193956 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 24 11:37:17.195701 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 24 11:37:17.239000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.239000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.283000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.263165 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 24 11:37:17.287397 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 24 11:37:17.288233 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 24 11:37:17.304000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.304000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.308201 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 24 11:37:17.320612 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 24 11:37:17.358000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.358000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.360563 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 24 11:37:17.361545 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 24 11:37:17.379000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.379000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:37:17.391173 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 24 11:37:17.391598 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 24 11:37:17.393000 audit: BPF prog-id=56 op=LOAD Jan 24 11:37:17.399033 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 24 11:37:17.525138 augenrules[1481]: No rules Jan 24 11:37:17.523000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Jan 24 11:37:17.523000 audit[1481]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffeca5c0930 a2=420 a3=0 items=0 ppid=1444 pid=1481 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:37:17.523000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 24 11:37:17.552561 systemd[1]: audit-rules.service: Deactivated successfully. Jan 24 11:37:17.553922 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 24 11:37:17.600597 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 24 11:37:17.645136 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 24 11:37:17.680664 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 24 11:37:17.696081 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 24 11:37:18.594662 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jan 24 11:37:18.635269 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 24 11:37:18.666560 systemd[1]: Reached target time-set.target - System Time Set. Jan 24 11:37:19.220179 systemd-networkd[1501]: lo: Link UP Jan 24 11:37:19.278072 systemd-networkd[1501]: lo: Gained carrier Jan 24 11:37:19.307272 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 24 11:37:19.331580 systemd[1]: Reached target network.target - Network. Jan 24 11:37:19.369980 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Jan 24 11:37:19.412259 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 24 11:37:20.287114 kernel: mousedev: PS/2 mouse device common for all mice Jan 24 11:37:20.332365 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Jan 24 11:37:20.405065 systemd-networkd[1501]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 24 11:37:20.405705 systemd-networkd[1501]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 24 11:37:20.457190 systemd-networkd[1501]: eth0: Link UP Jan 24 11:37:20.457681 systemd-networkd[1501]: eth0: Gained carrier Jan 24 11:37:20.457900 systemd-networkd[1501]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 24 11:37:20.682239 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 24 11:37:20.953629 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 24 11:37:21.070338 systemd-networkd[1501]: eth0: DHCPv4 address 10.0.0.26/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 24 11:37:21.080566 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jan 24 11:37:21.078616 systemd-timesyncd[1478]: Network configuration changed, trying to establish connection. Jan 24 11:37:21.657781 systemd-resolved[1287]: Clock change detected. Flushing caches. Jan 24 11:37:21.659094 systemd-timesyncd[1478]: Contacted time server 10.0.0.1:123 (10.0.0.1). Jan 24 11:37:21.662413 systemd-timesyncd[1478]: Initial clock synchronization to Sat 2026-01-24 11:37:21.657448 UTC. Jan 24 11:37:21.746089 kernel: ACPI: button: Power Button [PWRF] Jan 24 11:37:21.795032 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 24 11:37:22.083376 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Jan 24 11:37:22.101840 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Jan 24 11:37:22.133135 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Jan 24 11:37:22.162417 systemd-networkd[1501]: eth0: Gained IPv6LL Jan 24 11:37:22.188407 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 24 11:37:22.232892 systemd[1]: Reached target network-online.target - Network is Online. Jan 24 11:37:23.444593 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 24 11:37:23.728541 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 24 11:37:23.729758 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:37:23.770858 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 24 11:37:25.675083 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:37:25.696427 ldconfig[1446]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 24 11:37:25.700536 kernel: kvm_amd: TSC scaling supported Jan 24 11:37:25.700731 kernel: kvm_amd: Nested Virtualization enabled Jan 24 11:37:25.732583 kernel: kvm_amd: Nested Paging enabled Jan 24 11:37:25.733563 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Jan 24 11:37:25.733766 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 24 11:37:25.747504 kernel: kvm_amd: PMU virtualization is disabled Jan 24 11:37:25.788729 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 24 11:37:26.640663 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 24 11:37:26.663517 systemd[1]: Reached target sysinit.target - System Initialization. Jan 24 11:37:26.678422 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 24 11:37:26.692457 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 24 11:37:26.721702 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Jan 24 11:37:26.735802 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 24 11:37:26.747849 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 24 11:37:26.764666 systemd[1]: Started systemd-sysupdate-reboot.timer - Reboot Automatically After System Update. Jan 24 11:37:26.781482 systemd[1]: Started systemd-sysupdate.timer - Automatic System Update. Jan 24 11:37:26.820817 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 24 11:37:26.835356 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 24 11:37:26.835487 systemd[1]: Reached target paths.target - Path Units. Jan 24 11:37:26.847168 systemd[1]: Reached target timers.target - Timer Units. Jan 24 11:37:26.874635 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 24 11:37:26.927638 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 24 11:37:26.999091 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Jan 24 11:37:27.093809 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Jan 24 11:37:27.122667 systemd[1]: Reached target ssh-access.target - SSH Access Available. Jan 24 11:37:27.471443 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 24 11:37:27.491494 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Jan 24 11:37:27.557172 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 24 11:37:27.649860 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 24 11:37:27.677041 systemd[1]: Reached target sockets.target - Socket Units. Jan 24 11:37:27.691647 systemd[1]: Reached target basic.target - Basic System. Jan 24 11:37:27.724846 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 24 11:37:27.728676 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 24 11:37:27.735469 systemd[1]: Starting containerd.service - containerd container runtime... Jan 24 11:37:27.767760 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jan 24 11:37:27.841388 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 24 11:37:27.884643 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 24 11:37:27.966752 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 24 11:37:28.030681 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 24 11:37:28.063773 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 24 11:37:28.082638 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Jan 24 11:37:28.143592 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:37:28.167616 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 24 11:37:28.181439 jq[1562]: false Jan 24 11:37:28.187119 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 24 11:37:28.193446 kernel: EDAC MC: Ver: 3.0.0 Jan 24 11:37:28.354429 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 24 11:37:28.772913 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 24 11:37:28.789795 extend-filesystems[1563]: Found /dev/vda6 Jan 24 11:37:28.831608 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 24 11:37:28.836833 extend-filesystems[1563]: Found /dev/vda9 Jan 24 11:37:28.868775 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 24 11:37:28.902095 extend-filesystems[1563]: Checking size of /dev/vda9 Jan 24 11:37:28.883697 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 24 11:37:28.900626 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 24 11:37:28.906690 systemd[1]: Starting update-engine.service - Update Engine... Jan 24 11:37:29.083471 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 24 11:37:29.483339 extend-filesystems[1563]: Resized partition /dev/vda9 Jan 24 11:37:29.520507 google_oslogin_nss_cache[1564]: oslogin_cache_refresh[1564]: Refreshing passwd entry cache Jan 24 11:37:29.519775 oslogin_cache_refresh[1564]: Refreshing passwd entry cache Jan 24 11:37:29.542697 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 24 11:37:29.563059 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 24 11:37:29.625641 extend-filesystems[1595]: resize2fs 1.47.3 (8-Jul-2025) Jan 24 11:37:29.566556 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 24 11:37:29.606930 systemd[1]: motdgen.service: Deactivated successfully. Jan 24 11:37:29.653494 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 24 11:37:29.772457 google_oslogin_nss_cache[1564]: oslogin_cache_refresh[1564]: Failure getting users, quitting Jan 24 11:37:29.772457 google_oslogin_nss_cache[1564]: oslogin_cache_refresh[1564]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 24 11:37:29.772457 google_oslogin_nss_cache[1564]: oslogin_cache_refresh[1564]: Refreshing group entry cache Jan 24 11:37:29.736822 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 24 11:37:29.738497 oslogin_cache_refresh[1564]: Failure getting users, quitting Jan 24 11:37:29.772727 jq[1588]: true Jan 24 11:37:29.739539 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 24 11:37:29.738633 oslogin_cache_refresh[1564]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 24 11:37:29.826091 google_oslogin_nss_cache[1564]: oslogin_cache_refresh[1564]: Failure getting groups, quitting Jan 24 11:37:29.826091 google_oslogin_nss_cache[1564]: oslogin_cache_refresh[1564]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 24 11:37:29.738820 oslogin_cache_refresh[1564]: Refreshing group entry cache Jan 24 11:37:29.794129 oslogin_cache_refresh[1564]: Failure getting groups, quitting Jan 24 11:37:29.794156 oslogin_cache_refresh[1564]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 24 11:37:29.869666 kernel: EXT4-fs (vda9): resizing filesystem from 456704 to 1784827 blocks Jan 24 11:37:29.885738 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Jan 24 11:37:29.887494 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Jan 24 11:37:29.982463 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 24 11:37:30.578504 update_engine[1582]: I20260124 11:37:30.484858 1582 main.cc:92] Flatcar Update Engine starting Jan 24 11:37:30.624148 jq[1605]: true Jan 24 11:37:30.823525 kernel: EXT4-fs (vda9): resized filesystem to 1784827 Jan 24 11:37:30.586384 systemd[1]: coreos-metadata.service: Deactivated successfully. Jan 24 11:37:30.587047 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jan 24 11:37:30.793350 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 24 11:37:30.826668 extend-filesystems[1595]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jan 24 11:37:30.826668 extend-filesystems[1595]: old_desc_blocks = 1, new_desc_blocks = 1 Jan 24 11:37:30.826668 extend-filesystems[1595]: The filesystem on /dev/vda9 is now 1784827 (4k) blocks long. Jan 24 11:37:30.864542 systemd-logind[1579]: Watching system buttons on /dev/input/event2 (Power Button) Jan 24 11:37:30.975583 extend-filesystems[1563]: Resized filesystem in /dev/vda9 Jan 24 11:37:30.864590 systemd-logind[1579]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 24 11:37:30.865629 systemd-logind[1579]: New seat seat0. Jan 24 11:37:30.872051 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 24 11:37:30.873636 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 24 11:37:30.955762 systemd[1]: Started systemd-logind.service - User Login Management. Jan 24 11:37:31.686942 tar[1601]: linux-amd64/LICENSE Jan 24 11:37:31.696149 dbus-daemon[1560]: [system] SELinux support is enabled Jan 24 11:37:31.699619 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 24 11:37:31.754471 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 24 11:37:31.771827 tar[1601]: linux-amd64/helm Jan 24 11:37:31.754524 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 24 11:37:31.780821 update_engine[1582]: I20260124 11:37:31.780588 1582 update_check_scheduler.cc:74] Next update check in 11m2s Jan 24 11:37:31.847788 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 24 11:37:31.847843 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 24 11:37:31.964773 systemd[1]: Started update-engine.service - Update Engine. Jan 24 11:37:32.092760 dbus-daemon[1560]: [system] Successfully activated service 'org.freedesktop.systemd1' Jan 24 11:37:32.389390 bash[1646]: Updated "/home/core/.ssh/authorized_keys" Jan 24 11:37:32.933570 sshd_keygen[1622]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 24 11:37:33.742812 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 24 11:37:33.764405 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 24 11:37:33.846401 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jan 24 11:37:33.876866 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 24 11:37:33.972146 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 24 11:37:34.053638 systemd[1]: Started sshd@0-10.0.0.26:22-10.0.0.1:41582.service - OpenSSH per-connection server daemon (10.0.0.1:41582). Jan 24 11:37:35.426612 systemd[1]: issuegen.service: Deactivated successfully. Jan 24 11:37:35.427694 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 24 11:37:35.718543 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 24 11:37:37.931392 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 24 11:37:37.979467 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 24 11:37:38.020669 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 24 11:37:38.037904 systemd[1]: Reached target getty.target - Login Prompts. Jan 24 11:37:38.191470 locksmithd[1647]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 24 11:37:40.282505 sshd[1658]: Accepted publickey for core from 10.0.0.1 port 41582 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:37:40.301119 sshd-session[1658]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:37:40.825709 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 24 11:37:40.876431 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 24 11:37:41.326475 systemd-logind[1579]: New session 1 of user core. Jan 24 11:37:42.456786 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 24 11:37:42.489653 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 24 11:37:43.333994 (systemd)[1682]: pam_unix(systemd-user:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:37:43.353175 systemd-logind[1579]: New session 2 of user core. Jan 24 11:37:44.739667 containerd[1625]: time="2026-01-24T11:37:44Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Jan 24 11:37:44.826796 containerd[1625]: time="2026-01-24T11:37:44.812461414Z" level=info msg="starting containerd" revision=fcd43222d6b07379a4be9786bda52438f0dd16a1 version=v2.1.5 Jan 24 11:37:45.845932 containerd[1625]: time="2026-01-24T11:37:45.845667922Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="876.886µs" Jan 24 11:37:45.853848 containerd[1625]: time="2026-01-24T11:37:45.853629065Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Jan 24 11:37:45.855770 containerd[1625]: time="2026-01-24T11:37:45.855733374Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Jan 24 11:37:45.884652 containerd[1625]: time="2026-01-24T11:37:45.864932368Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Jan 24 11:37:45.933678 containerd[1625]: time="2026-01-24T11:37:45.925882728Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Jan 24 11:37:45.997683 containerd[1625]: time="2026-01-24T11:37:45.994146403Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 24 11:37:46.084956 containerd[1625]: time="2026-01-24T11:37:46.083814347Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 24 11:37:46.365811 containerd[1625]: time="2026-01-24T11:37:46.223196515Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 24 11:37:46.226684 systemd[1682]: Queued start job for default target default.target. Jan 24 11:37:46.628776 systemd[1682]: Created slice app.slice - User Application Slice. Jan 24 11:37:46.630685 systemd[1682]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. Jan 24 11:37:46.630721 systemd[1682]: Reached target paths.target - Paths. Jan 24 11:37:46.631880 systemd[1682]: Reached target timers.target - Timers. Jan 24 11:37:46.644155 systemd[1682]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 24 11:37:46.664484 systemd[1682]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... Jan 24 11:37:46.673575 containerd[1625]: time="2026-01-24T11:37:46.671987921Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 24 11:37:46.673728 containerd[1625]: time="2026-01-24T11:37:46.673693766Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 24 11:37:46.674994 containerd[1625]: time="2026-01-24T11:37:46.674857178Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 24 11:37:46.675399 containerd[1625]: time="2026-01-24T11:37:46.675371568Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 24 11:37:46.678743 containerd[1625]: time="2026-01-24T11:37:46.678714799Z" level=info msg="skip loading plugin" error="EROFS unsupported, please `modprobe erofs`: skip plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 24 11:37:46.678842 containerd[1625]: time="2026-01-24T11:37:46.678813093Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Jan 24 11:37:46.681373 containerd[1625]: time="2026-01-24T11:37:46.680456521Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Jan 24 11:37:46.686946 containerd[1625]: time="2026-01-24T11:37:46.686906744Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 24 11:37:46.688480 containerd[1625]: time="2026-01-24T11:37:46.688161075Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 24 11:37:46.688587 containerd[1625]: time="2026-01-24T11:37:46.688565330Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Jan 24 11:37:46.689752 containerd[1625]: time="2026-01-24T11:37:46.689414225Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Jan 24 11:37:46.721819 containerd[1625]: time="2026-01-24T11:37:46.721768028Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Jan 24 11:37:46.722718 containerd[1625]: time="2026-01-24T11:37:46.722681894Z" level=info msg="metadata content store policy set" policy=shared Jan 24 11:37:47.685960 containerd[1625]: time="2026-01-24T11:37:47.685505346Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Jan 24 11:37:47.695940 containerd[1625]: time="2026-01-24T11:37:47.695906193Z" level=info msg="loading plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 24 11:37:47.701013 containerd[1625]: time="2026-01-24T11:37:47.700968174Z" level=info msg="skip loading plugin" error="could not find mkfs.erofs: exec: \"mkfs.erofs\": executable file not found in $PATH: skip plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 24 11:37:47.701764 containerd[1625]: time="2026-01-24T11:37:47.701735246Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Jan 24 11:37:47.702719 containerd[1625]: time="2026-01-24T11:37:47.702689668Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Jan 24 11:37:47.729784 containerd[1625]: time="2026-01-24T11:37:47.729739368Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Jan 24 11:37:47.730607 containerd[1625]: time="2026-01-24T11:37:47.730577563Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Jan 24 11:37:47.730718 containerd[1625]: time="2026-01-24T11:37:47.730692117Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Jan 24 11:37:47.731373 containerd[1625]: time="2026-01-24T11:37:47.731169208Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Jan 24 11:37:47.731562 containerd[1625]: time="2026-01-24T11:37:47.731540381Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Jan 24 11:37:47.731831 containerd[1625]: time="2026-01-24T11:37:47.731807940Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Jan 24 11:37:47.731904 containerd[1625]: time="2026-01-24T11:37:47.731889442Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Jan 24 11:37:47.732559 containerd[1625]: time="2026-01-24T11:37:47.732533314Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Jan 24 11:37:47.732752 containerd[1625]: time="2026-01-24T11:37:47.732729210Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Jan 24 11:37:47.737746 containerd[1625]: time="2026-01-24T11:37:47.737716601Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Jan 24 11:37:47.739770 containerd[1625]: time="2026-01-24T11:37:47.739742353Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Jan 24 11:37:47.739874 containerd[1625]: time="2026-01-24T11:37:47.739853541Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Jan 24 11:37:47.740627 containerd[1625]: time="2026-01-24T11:37:47.740596858Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Jan 24 11:37:47.740719 containerd[1625]: time="2026-01-24T11:37:47.740697507Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Jan 24 11:37:47.741741 containerd[1625]: time="2026-01-24T11:37:47.740795940Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Jan 24 11:37:47.741854 containerd[1625]: time="2026-01-24T11:37:47.741832155Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Jan 24 11:37:47.742665 containerd[1625]: time="2026-01-24T11:37:47.742639321Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Jan 24 11:37:47.742979 containerd[1625]: time="2026-01-24T11:37:47.742956383Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Jan 24 11:37:47.743420 containerd[1625]: time="2026-01-24T11:37:47.743395984Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Jan 24 11:37:47.743504 containerd[1625]: time="2026-01-24T11:37:47.743483728Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Jan 24 11:37:47.744012 containerd[1625]: time="2026-01-24T11:37:47.743986196Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Jan 24 11:37:47.750873 containerd[1625]: time="2026-01-24T11:37:47.750837136Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Jan 24 11:37:47.753024 containerd[1625]: time="2026-01-24T11:37:47.752996478Z" level=info msg="Start snapshots syncer" Jan 24 11:37:47.761917 containerd[1625]: time="2026-01-24T11:37:47.754167093Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Jan 24 11:37:47.777336 containerd[1625]: time="2026-01-24T11:37:47.776945141Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"cgroupWritable\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"\",\"binDirs\":[\"/opt/cni/bin\"],\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogLineSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Jan 24 11:37:47.779488 containerd[1625]: time="2026-01-24T11:37:47.778137096Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Jan 24 11:37:47.779488 containerd[1625]: time="2026-01-24T11:37:47.779144657Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.779808427Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.779852078Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.779872908Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.779887845Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.780006637Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.780144274Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.780368082Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.780475803Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.780497333Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Jan 24 11:37:47.780749 containerd[1625]: time="2026-01-24T11:37:47.780742260Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 24 11:37:47.781664 containerd[1625]: time="2026-01-24T11:37:47.780768038Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 24 11:37:47.781664 containerd[1625]: time="2026-01-24T11:37:47.780782225Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 24 11:37:47.781664 containerd[1625]: time="2026-01-24T11:37:47.780795259Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 24 11:37:47.781664 containerd[1625]: time="2026-01-24T11:37:47.780807001Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Jan 24 11:37:47.781664 containerd[1625]: time="2026-01-24T11:37:47.780827840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Jan 24 11:37:47.781664 containerd[1625]: time="2026-01-24T11:37:47.781132780Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Jan 24 11:37:47.783325 containerd[1625]: time="2026-01-24T11:37:47.782145700Z" level=info msg="runtime interface created" Jan 24 11:37:47.786570 containerd[1625]: time="2026-01-24T11:37:47.786390075Z" level=info msg="created NRI interface" Jan 24 11:37:47.786570 containerd[1625]: time="2026-01-24T11:37:47.786525066Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Jan 24 11:37:47.786678 containerd[1625]: time="2026-01-24T11:37:47.786595258Z" level=info msg="Connect containerd service" Jan 24 11:37:47.787522 containerd[1625]: time="2026-01-24T11:37:47.786925414Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 24 11:37:47.824954 systemd[1682]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 24 11:37:47.825631 systemd[1682]: Reached target sockets.target - Sockets. Jan 24 11:37:48.177868 containerd[1625]: time="2026-01-24T11:37:48.143211478Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 24 11:37:48.240635 systemd[1682]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. Jan 24 11:37:48.244816 systemd[1682]: Reached target basic.target - Basic System. Jan 24 11:37:48.245554 systemd[1682]: Reached target default.target - Main User Target. Jan 24 11:37:48.245728 systemd[1682]: Startup finished in 4.660s. Jan 24 11:37:48.247582 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 24 11:37:48.283542 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 24 11:37:48.434757 systemd[1]: Started sshd@1-10.0.0.26:22-10.0.0.1:34568.service - OpenSSH per-connection server daemon (10.0.0.1:34568). Jan 24 11:37:50.140649 tar[1601]: linux-amd64/README.md Jan 24 11:37:50.886992 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 24 11:37:50.894382 sshd[1705]: Accepted publickey for core from 10.0.0.1 port 34568 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:37:50.935796 sshd-session[1705]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:37:50.971604 systemd-logind[1579]: New session 3 of user core. Jan 24 11:37:51.003703 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 24 11:37:52.166744 sshd[1716]: Connection closed by 10.0.0.1 port 34568 Jan 24 11:37:52.195893 sshd-session[1705]: pam_unix(sshd:session): session closed for user core Jan 24 11:37:52.443595 systemd[1]: sshd@1-10.0.0.26:22-10.0.0.1:34568.service: Deactivated successfully. Jan 24 11:37:52.790862 systemd[1]: session-3.scope: Deactivated successfully. Jan 24 11:37:52.834381 systemd-logind[1579]: Session 3 logged out. Waiting for processes to exit. Jan 24 11:37:52.837442 systemd[1]: Started sshd@2-10.0.0.26:22-10.0.0.1:34588.service - OpenSSH per-connection server daemon (10.0.0.1:34588). Jan 24 11:37:52.840793 systemd-logind[1579]: Removed session 3. Jan 24 11:37:53.329711 containerd[1625]: time="2026-01-24T11:37:53.325016005Z" level=info msg="Start subscribing containerd event" Jan 24 11:37:53.358720 containerd[1625]: time="2026-01-24T11:37:53.354863861Z" level=info msg="Start recovering state" Jan 24 11:37:53.362899 containerd[1625]: time="2026-01-24T11:37:53.362644580Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 24 11:37:53.364496 containerd[1625]: time="2026-01-24T11:37:53.364464037Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 24 11:37:53.369587 containerd[1625]: time="2026-01-24T11:37:53.369554730Z" level=info msg="Start event monitor" Jan 24 11:37:53.370548 containerd[1625]: time="2026-01-24T11:37:53.370523960Z" level=info msg="Start cni network conf syncer for default" Jan 24 11:37:53.371389 containerd[1625]: time="2026-01-24T11:37:53.371173814Z" level=info msg="Start streaming server" Jan 24 11:37:53.371678 containerd[1625]: time="2026-01-24T11:37:53.371652578Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Jan 24 11:37:53.371760 containerd[1625]: time="2026-01-24T11:37:53.371742586Z" level=info msg="runtime interface starting up..." Jan 24 11:37:53.371831 containerd[1625]: time="2026-01-24T11:37:53.371815211Z" level=info msg="starting plugins..." Jan 24 11:37:53.372040 containerd[1625]: time="2026-01-24T11:37:53.372021306Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Jan 24 11:37:53.420046 systemd[1]: Started containerd.service - containerd container runtime. Jan 24 11:37:53.423507 containerd[1625]: time="2026-01-24T11:37:53.423457313Z" level=info msg="containerd successfully booted in 8.804281s" Jan 24 11:37:53.607787 sshd[1730]: Accepted publickey for core from 10.0.0.1 port 34588 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:37:53.827040 sshd-session[1730]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:37:53.872010 systemd-logind[1579]: New session 4 of user core. Jan 24 11:37:53.891635 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 24 11:37:54.036839 sshd[1734]: Connection closed by 10.0.0.1 port 34588 Jan 24 11:37:54.039987 sshd-session[1730]: pam_unix(sshd:session): session closed for user core Jan 24 11:37:54.050966 systemd[1]: sshd@2-10.0.0.26:22-10.0.0.1:34588.service: Deactivated successfully. Jan 24 11:37:54.058464 systemd[1]: session-4.scope: Deactivated successfully. Jan 24 11:37:54.081794 systemd-logind[1579]: Session 4 logged out. Waiting for processes to exit. Jan 24 11:37:54.090593 systemd-logind[1579]: Removed session 4. Jan 24 11:38:02.393658 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:38:02.398587 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 24 11:38:02.408901 systemd[1]: Startup finished in 28.615s (kernel) + 32.498s (initrd) + 1min 14.331s (userspace) = 2min 15.445s. Jan 24 11:38:02.478040 (kubelet)[1744]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:38:04.148571 systemd[1]: Started sshd@3-10.0.0.26:22-10.0.0.1:38204.service - OpenSSH per-connection server daemon (10.0.0.1:38204). Jan 24 11:38:06.491955 sshd[1746]: Accepted publickey for core from 10.0.0.1 port 38204 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:38:06.547951 sshd-session[1746]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:38:06.902965 systemd-logind[1579]: New session 5 of user core. Jan 24 11:38:06.924109 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 24 11:38:07.245136 sshd[1755]: Connection closed by 10.0.0.1 port 38204 Jan 24 11:38:07.250864 sshd-session[1746]: pam_unix(sshd:session): session closed for user core Jan 24 11:38:07.273422 systemd[1]: sshd@3-10.0.0.26:22-10.0.0.1:38204.service: Deactivated successfully. Jan 24 11:38:07.280803 systemd[1]: session-5.scope: Deactivated successfully. Jan 24 11:38:07.298537 systemd-logind[1579]: Session 5 logged out. Waiting for processes to exit. Jan 24 11:38:07.315544 systemd[1]: Started sshd@4-10.0.0.26:22-10.0.0.1:38230.service - OpenSSH per-connection server daemon (10.0.0.1:38230). Jan 24 11:38:07.320545 systemd-logind[1579]: Removed session 5. Jan 24 11:38:08.572766 sshd[1761]: Accepted publickey for core from 10.0.0.1 port 38230 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:38:08.578040 sshd-session[1761]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:38:09.191846 systemd-logind[1579]: New session 6 of user core. Jan 24 11:38:09.498030 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 24 11:38:09.886921 sshd[1765]: Connection closed by 10.0.0.1 port 38230 Jan 24 11:38:09.893509 sshd-session[1761]: pam_unix(sshd:session): session closed for user core Jan 24 11:38:10.183577 systemd[1]: sshd@4-10.0.0.26:22-10.0.0.1:38230.service: Deactivated successfully. Jan 24 11:38:10.244000 systemd[1]: session-6.scope: Deactivated successfully. Jan 24 11:38:10.342825 systemd-logind[1579]: Session 6 logged out. Waiting for processes to exit. Jan 24 11:38:10.655817 systemd[1]: Started sshd@5-10.0.0.26:22-10.0.0.1:38246.service - OpenSSH per-connection server daemon (10.0.0.1:38246). Jan 24 11:38:10.768842 systemd-logind[1579]: Removed session 6. Jan 24 11:38:12.134807 sshd[1771]: Accepted publickey for core from 10.0.0.1 port 38246 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:38:12.163962 sshd-session[1771]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:38:12.741867 systemd-logind[1579]: New session 7 of user core. Jan 24 11:38:12.791699 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 24 11:38:13.317850 sshd[1776]: Connection closed by 10.0.0.1 port 38246 Jan 24 11:38:13.324782 sshd-session[1771]: pam_unix(sshd:session): session closed for user core Jan 24 11:38:13.495678 systemd[1]: sshd@5-10.0.0.26:22-10.0.0.1:38246.service: Deactivated successfully. Jan 24 11:38:13.512865 systemd[1]: session-7.scope: Deactivated successfully. Jan 24 11:38:13.516932 systemd-logind[1579]: Session 7 logged out. Waiting for processes to exit. Jan 24 11:38:13.534522 systemd[1]: Started sshd@6-10.0.0.26:22-10.0.0.1:51828.service - OpenSSH per-connection server daemon (10.0.0.1:51828). Jan 24 11:38:13.539030 systemd-logind[1579]: Removed session 7. Jan 24 11:38:14.320442 sshd[1782]: Accepted publickey for core from 10.0.0.1 port 51828 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:38:14.335949 sshd-session[1782]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:38:14.740126 systemd-logind[1579]: New session 8 of user core. Jan 24 11:38:14.842452 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 24 11:38:15.548156 sudo[1787]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 24 11:38:15.550113 sudo[1787]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 24 11:38:15.588828 sudo[1787]: pam_unix(sudo:session): session closed for user root Jan 24 11:38:15.597440 sshd[1786]: Connection closed by 10.0.0.1 port 51828 Jan 24 11:38:15.622460 sshd-session[1782]: pam_unix(sshd:session): session closed for user core Jan 24 11:38:15.691356 systemd[1]: sshd@6-10.0.0.26:22-10.0.0.1:51828.service: Deactivated successfully. Jan 24 11:38:15.894085 systemd[1]: session-8.scope: Deactivated successfully. Jan 24 11:38:15.912884 systemd-logind[1579]: Session 8 logged out. Waiting for processes to exit. Jan 24 11:38:15.922407 systemd[1]: Started sshd@7-10.0.0.26:22-10.0.0.1:51836.service - OpenSSH per-connection server daemon (10.0.0.1:51836). Jan 24 11:38:15.933758 systemd-logind[1579]: Removed session 8. Jan 24 11:38:16.431136 kubelet[1744]: E0124 11:38:16.430450 1744 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:38:16.435603 sshd[1794]: Accepted publickey for core from 10.0.0.1 port 51836 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:38:16.441845 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:38:16.442893 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:38:16.443962 systemd[1]: kubelet.service: Consumed 20.633s CPU time, 259.4M memory peak. Jan 24 11:38:16.450685 sshd-session[1794]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:38:16.470500 systemd-logind[1579]: New session 9 of user core. Jan 24 11:38:16.492876 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 24 11:38:16.741911 sudo[1801]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 24 11:38:16.742990 sudo[1801]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 24 11:38:16.769774 sudo[1801]: pam_unix(sudo:session): session closed for user root Jan 24 11:38:16.821791 sudo[1800]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jan 24 11:38:16.822753 sudo[1800]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 24 11:38:16.854373 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 24 11:38:16.968833 update_engine[1582]: I20260124 11:38:16.966913 1582 update_attempter.cc:509] Updating boot flags... Jan 24 11:38:17.126000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 24 11:38:17.135864 augenrules[1831]: No rules Jan 24 11:38:17.137811 kernel: kauditd_printk_skb: 69 callbacks suppressed Jan 24 11:38:17.137890 kernel: audit: type=1305 audit(1769254697.126:218): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 24 11:38:17.157160 systemd[1]: audit-rules.service: Deactivated successfully. Jan 24 11:38:17.126000 audit[1831]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc5453da20 a2=420 a3=0 items=0 ppid=1806 pid=1831 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:17.168595 sudo[1800]: pam_unix(sudo:session): session closed for user root Jan 24 11:38:17.158883 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 24 11:38:17.179432 sshd[1799]: Connection closed by 10.0.0.1 port 51836 Jan 24 11:38:17.190895 sshd-session[1794]: pam_unix(sshd:session): session closed for user core Jan 24 11:38:17.203834 kernel: audit: type=1300 audit(1769254697.126:218): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc5453da20 a2=420 a3=0 items=0 ppid=1806 pid=1831 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:17.203938 kernel: audit: type=1327 audit(1769254697.126:218): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 24 11:38:17.126000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 24 11:38:17.241811 kernel: audit: type=1130 audit(1769254697.157:219): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.243328 kernel: audit: type=1131 audit(1769254697.157:220): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.157000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.157000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.268687 kernel: audit: type=1106 audit(1769254697.163:221): pid=1800 uid=500 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.163000 audit[1800]: USER_END pid=1800 uid=500 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.280372 systemd[1]: sshd@7-10.0.0.26:22-10.0.0.1:51836.service: Deactivated successfully. Jan 24 11:38:17.287019 systemd[1]: session-9.scope: Deactivated successfully. Jan 24 11:38:17.301464 kernel: audit: type=1104 audit(1769254697.163:222): pid=1800 uid=500 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.163000 audit[1800]: CRED_DISP pid=1800 uid=500 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.317614 systemd-logind[1579]: Session 9 logged out. Waiting for processes to exit. Jan 24 11:38:17.193000 audit[1794]: USER_END pid=1794 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:38:17.359606 kernel: audit: type=1106 audit(1769254697.193:223): pid=1794 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:38:17.359749 kernel: audit: type=1104 audit(1769254697.193:224): pid=1794 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:38:17.193000 audit[1794]: CRED_DISP pid=1794 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:38:17.360732 systemd-logind[1579]: Removed session 9. Jan 24 11:38:17.366800 systemd[1]: Started sshd@8-10.0.0.26:22-10.0.0.1:51844.service - OpenSSH per-connection server daemon (10.0.0.1:51844). Jan 24 11:38:17.280000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.26:22-10.0.0.1:51836 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.365000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.26:22-10.0.0.1:51844 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.384405 kernel: audit: type=1131 audit(1769254697.280:225): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.26:22-10.0.0.1:51836 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:17.758000 audit[1849]: USER_ACCT pid=1849 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:38:17.766747 sshd[1849]: Accepted publickey for core from 10.0.0.1 port 51844 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:38:17.772098 sshd-session[1849]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:38:17.769000 audit[1849]: CRED_ACQ pid=1849 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:38:17.769000 audit[1849]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc3518ec90 a2=3 a3=0 items=0 ppid=1 pid=1849 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:17.769000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:38:17.828036 systemd-logind[1579]: New session 10 of user core. Jan 24 11:38:17.840674 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 24 11:38:17.853000 audit[1849]: USER_START pid=1849 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:38:17.859000 audit[1853]: CRED_ACQ pid=1853 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:38:18.086000 audit[1854]: USER_ACCT pid=1854 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:38:18.115000 audit[1854]: CRED_REFR pid=1854 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:38:18.115746 sudo[1854]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 24 11:38:18.116000 audit[1854]: USER_START pid=1854 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:38:18.117945 sudo[1854]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 24 11:38:25.246145 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 24 11:38:25.286956 (dockerd)[1876]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 24 11:38:26.688360 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 24 11:38:26.725779 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:38:30.502109 dockerd[1876]: time="2026-01-24T11:38:30.487593913Z" level=info msg="Starting up" Jan 24 11:38:30.532361 dockerd[1876]: time="2026-01-24T11:38:30.531456578Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Jan 24 11:38:31.032388 dockerd[1876]: time="2026-01-24T11:38:31.031742119Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Jan 24 11:38:31.426000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:31.427620 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:38:31.439559 kernel: kauditd_printk_skb: 11 callbacks suppressed Jan 24 11:38:31.439722 kernel: audit: type=1130 audit(1769254711.426:235): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:31.625944 (kubelet)[1909]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:38:31.866987 dockerd[1876]: time="2026-01-24T11:38:31.864442568Z" level=info msg="Loading containers: start." Jan 24 11:38:31.973371 kernel: Initializing XFRM netlink socket Jan 24 11:38:32.283440 kubelet[1909]: E0124 11:38:32.283378 1909 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:38:32.296105 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:38:32.297791 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:38:32.303000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:38:32.304826 systemd[1]: kubelet.service: Consumed 3.082s CPU time, 108.3M memory peak. Jan 24 11:38:32.335333 kernel: audit: type=1131 audit(1769254712.303:236): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:38:32.789000 audit[1947]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=1947 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:32.789000 audit[1947]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffd10e6a3f0 a2=0 a3=0 items=0 ppid=1876 pid=1947 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:32.877418 kernel: audit: type=1325 audit(1769254712.789:237): table=nat:2 family=2 entries=2 op=nft_register_chain pid=1947 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:32.881626 kernel: audit: type=1300 audit(1769254712.789:237): arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffd10e6a3f0 a2=0 a3=0 items=0 ppid=1876 pid=1947 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:32.881725 kernel: audit: type=1327 audit(1769254712.789:237): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 24 11:38:32.789000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 24 11:38:32.838000 audit[1949]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=1949 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:32.924881 kernel: audit: type=1325 audit(1769254712.838:238): table=filter:3 family=2 entries=2 op=nft_register_chain pid=1949 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:32.931719 kernel: audit: type=1300 audit(1769254712.838:238): arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffcd9498770 a2=0 a3=0 items=0 ppid=1876 pid=1949 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:32.838000 audit[1949]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffcd9498770 a2=0 a3=0 items=0 ppid=1876 pid=1949 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:32.969789 kernel: audit: type=1327 audit(1769254712.838:238): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 24 11:38:32.838000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 24 11:38:32.860000 audit[1951]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=1951 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:32.860000 audit[1951]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc055b55e0 a2=0 a3=0 items=0 ppid=1876 pid=1951 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.046953 kernel: audit: type=1325 audit(1769254712.860:239): table=filter:4 family=2 entries=1 op=nft_register_chain pid=1951 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:33.047369 kernel: audit: type=1300 audit(1769254712.860:239): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc055b55e0 a2=0 a3=0 items=0 ppid=1876 pid=1951 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:32.860000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 24 11:38:32.882000 audit[1953]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=1953 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:32.882000 audit[1953]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff1621f710 a2=0 a3=0 items=0 ppid=1876 pid=1953 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:32.882000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 24 11:38:32.950000 audit[1955]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_chain pid=1955 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:32.950000 audit[1955]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffebcbb26d0 a2=0 a3=0 items=0 ppid=1876 pid=1955 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:32.950000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 24 11:38:32.979000 audit[1957]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_chain pid=1957 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:32.979000 audit[1957]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffc51815830 a2=0 a3=0 items=0 ppid=1876 pid=1957 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:32.979000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 24 11:38:33.016000 audit[1959]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=1959 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:33.016000 audit[1959]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffe0f009310 a2=0 a3=0 items=0 ppid=1876 pid=1959 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.016000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 24 11:38:33.032000 audit[1961]: NETFILTER_CFG table=nat:9 family=2 entries=2 op=nft_register_chain pid=1961 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:33.032000 audit[1961]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffd4badbc90 a2=0 a3=0 items=0 ppid=1876 pid=1961 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.032000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 24 11:38:33.244000 audit[1965]: NETFILTER_CFG table=nat:10 family=2 entries=2 op=nft_register_chain pid=1965 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:33.244000 audit[1965]: SYSCALL arch=c000003e syscall=46 success=yes exit=472 a0=3 a1=7ffccac86090 a2=0 a3=0 items=0 ppid=1876 pid=1965 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.244000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 Jan 24 11:38:33.291000 audit[1967]: NETFILTER_CFG table=filter:11 family=2 entries=2 op=nft_register_chain pid=1967 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:33.291000 audit[1967]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffd2d6e2ea0 a2=0 a3=0 items=0 ppid=1876 pid=1967 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.291000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 24 11:38:33.324000 audit[1969]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=1969 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:33.324000 audit[1969]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffdc0611990 a2=0 a3=0 items=0 ppid=1876 pid=1969 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.324000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 24 11:38:33.344000 audit[1971]: NETFILTER_CFG table=filter:13 family=2 entries=1 op=nft_register_rule pid=1971 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:33.344000 audit[1971]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7fffb4f566d0 a2=0 a3=0 items=0 ppid=1876 pid=1971 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.344000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 24 11:38:33.364000 audit[1973]: NETFILTER_CFG table=filter:14 family=2 entries=1 op=nft_register_rule pid=1973 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:33.364000 audit[1973]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7fff7a4c6ab0 a2=0 a3=0 items=0 ppid=1876 pid=1973 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.364000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 24 11:38:33.960000 audit[2003]: NETFILTER_CFG table=nat:15 family=10 entries=2 op=nft_register_chain pid=2003 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:33.960000 audit[2003]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffc98d5e6c0 a2=0 a3=0 items=0 ppid=1876 pid=2003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.960000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 24 11:38:33.989000 audit[2005]: NETFILTER_CFG table=filter:16 family=10 entries=2 op=nft_register_chain pid=2005 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:33.989000 audit[2005]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffedad30ec0 a2=0 a3=0 items=0 ppid=1876 pid=2005 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:33.989000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 24 11:38:34.035000 audit[2007]: NETFILTER_CFG table=filter:17 family=10 entries=1 op=nft_register_chain pid=2007 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.035000 audit[2007]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fffa2264310 a2=0 a3=0 items=0 ppid=1876 pid=2007 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.035000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 24 11:38:34.120000 audit[2009]: NETFILTER_CFG table=filter:18 family=10 entries=1 op=nft_register_chain pid=2009 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.120000 audit[2009]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffdf1bfeba0 a2=0 a3=0 items=0 ppid=1876 pid=2009 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.120000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 24 11:38:34.140000 audit[2011]: NETFILTER_CFG table=filter:19 family=10 entries=1 op=nft_register_chain pid=2011 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.140000 audit[2011]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffcc9964830 a2=0 a3=0 items=0 ppid=1876 pid=2011 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.140000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 24 11:38:34.158000 audit[2013]: NETFILTER_CFG table=filter:20 family=10 entries=1 op=nft_register_chain pid=2013 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.158000 audit[2013]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffc79d76130 a2=0 a3=0 items=0 ppid=1876 pid=2013 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.158000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 24 11:38:34.182000 audit[2015]: NETFILTER_CFG table=filter:21 family=10 entries=1 op=nft_register_chain pid=2015 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.182000 audit[2015]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffc4af84010 a2=0 a3=0 items=0 ppid=1876 pid=2015 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.182000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 24 11:38:34.220000 audit[2017]: NETFILTER_CFG table=nat:22 family=10 entries=2 op=nft_register_chain pid=2017 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.220000 audit[2017]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffdf6b3da70 a2=0 a3=0 items=0 ppid=1876 pid=2017 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.220000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 24 11:38:34.245000 audit[2019]: NETFILTER_CFG table=nat:23 family=10 entries=2 op=nft_register_chain pid=2019 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.245000 audit[2019]: SYSCALL arch=c000003e syscall=46 success=yes exit=484 a0=3 a1=7fff9b28d2b0 a2=0 a3=0 items=0 ppid=1876 pid=2019 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.245000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003A3A312F313238 Jan 24 11:38:34.261000 audit[2021]: NETFILTER_CFG table=filter:24 family=10 entries=2 op=nft_register_chain pid=2021 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.261000 audit[2021]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffcada136b0 a2=0 a3=0 items=0 ppid=1876 pid=2021 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.261000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 24 11:38:34.277000 audit[2023]: NETFILTER_CFG table=filter:25 family=10 entries=1 op=nft_register_rule pid=2023 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.277000 audit[2023]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffdc0f61680 a2=0 a3=0 items=0 ppid=1876 pid=2023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.277000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 24 11:38:34.296000 audit[2025]: NETFILTER_CFG table=filter:26 family=10 entries=1 op=nft_register_rule pid=2025 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.296000 audit[2025]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7fff784caf20 a2=0 a3=0 items=0 ppid=1876 pid=2025 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.296000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 24 11:38:34.318000 audit[2027]: NETFILTER_CFG table=filter:27 family=10 entries=1 op=nft_register_rule pid=2027 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.318000 audit[2027]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7fff14ad3fd0 a2=0 a3=0 items=0 ppid=1876 pid=2027 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.318000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 24 11:38:34.359000 audit[2032]: NETFILTER_CFG table=filter:28 family=2 entries=1 op=nft_register_chain pid=2032 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:34.359000 audit[2032]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe40cbb950 a2=0 a3=0 items=0 ppid=1876 pid=2032 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.359000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 24 11:38:34.372000 audit[2034]: NETFILTER_CFG table=filter:29 family=2 entries=1 op=nft_register_rule pid=2034 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:34.372000 audit[2034]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffed3345960 a2=0 a3=0 items=0 ppid=1876 pid=2034 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.372000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 24 11:38:34.386000 audit[2036]: NETFILTER_CFG table=filter:30 family=2 entries=1 op=nft_register_rule pid=2036 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:34.386000 audit[2036]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffdfc831ab0 a2=0 a3=0 items=0 ppid=1876 pid=2036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.386000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 24 11:38:34.426000 audit[2038]: NETFILTER_CFG table=filter:31 family=10 entries=1 op=nft_register_chain pid=2038 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.426000 audit[2038]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7fffdde24070 a2=0 a3=0 items=0 ppid=1876 pid=2038 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.426000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 24 11:38:34.455000 audit[2040]: NETFILTER_CFG table=filter:32 family=10 entries=1 op=nft_register_rule pid=2040 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.455000 audit[2040]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffe37c10e30 a2=0 a3=0 items=0 ppid=1876 pid=2040 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.455000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 24 11:38:34.479000 audit[2042]: NETFILTER_CFG table=filter:33 family=10 entries=1 op=nft_register_rule pid=2042 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:38:34.479000 audit[2042]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffc8e9af5d0 a2=0 a3=0 items=0 ppid=1876 pid=2042 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.479000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 24 11:38:34.642000 audit[2046]: NETFILTER_CFG table=nat:34 family=2 entries=2 op=nft_register_chain pid=2046 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:34.642000 audit[2046]: SYSCALL arch=c000003e syscall=46 success=yes exit=520 a0=3 a1=7ffca4d21ee0 a2=0 a3=0 items=0 ppid=1876 pid=2046 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.642000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 Jan 24 11:38:34.657000 audit[2048]: NETFILTER_CFG table=nat:35 family=2 entries=1 op=nft_register_rule pid=2048 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:34.657000 audit[2048]: SYSCALL arch=c000003e syscall=46 success=yes exit=288 a0=3 a1=7ffc896e2be0 a2=0 a3=0 items=0 ppid=1876 pid=2048 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.657000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E Jan 24 11:38:34.802000 audit[2056]: NETFILTER_CFG table=filter:36 family=2 entries=1 op=nft_register_rule pid=2056 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:34.802000 audit[2056]: SYSCALL arch=c000003e syscall=46 success=yes exit=300 a0=3 a1=7ffe28cf3f30 a2=0 a3=0 items=0 ppid=1876 pid=2056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.802000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D464F5257415244002D6900646F636B657230002D6A00414343455054 Jan 24 11:38:34.937000 audit[2062]: NETFILTER_CFG table=filter:37 family=2 entries=1 op=nft_register_rule pid=2062 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:34.937000 audit[2062]: SYSCALL arch=c000003e syscall=46 success=yes exit=376 a0=3 a1=7ffd62128730 a2=0 a3=0 items=0 ppid=1876 pid=2062 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:34.937000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45520000002D6900646F636B657230002D6F00646F636B657230002D6A0044524F50 Jan 24 11:38:35.029000 audit[2064]: NETFILTER_CFG table=filter:38 family=2 entries=1 op=nft_register_rule pid=2064 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:35.029000 audit[2064]: SYSCALL arch=c000003e syscall=46 success=yes exit=512 a0=3 a1=7fff5a2b1420 a2=0 a3=0 items=0 ppid=1876 pid=2064 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:35.029000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D4354002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 Jan 24 11:38:35.060000 audit[2066]: NETFILTER_CFG table=filter:39 family=2 entries=1 op=nft_register_rule pid=2066 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:35.060000 audit[2066]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffd616c5ba0 a2=0 a3=0 items=0 ppid=1876 pid=2066 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:35.060000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D425249444745002D6F00646F636B657230002D6A00444F434B4552 Jan 24 11:38:35.074000 audit[2068]: NETFILTER_CFG table=filter:40 family=2 entries=1 op=nft_register_rule pid=2068 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:35.074000 audit[2068]: SYSCALL arch=c000003e syscall=46 success=yes exit=428 a0=3 a1=7ffe00aa5260 a2=0 a3=0 items=0 ppid=1876 pid=2068 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:35.074000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 24 11:38:35.092000 audit[2070]: NETFILTER_CFG table=filter:41 family=2 entries=1 op=nft_register_rule pid=2070 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:38:35.092000 audit[2070]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffc8bfbc670 a2=0 a3=0 items=0 ppid=1876 pid=2070 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:38:35.092000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 Jan 24 11:38:35.102014 systemd-networkd[1501]: docker0: Link UP Jan 24 11:38:35.136383 dockerd[1876]: time="2026-01-24T11:38:35.135358458Z" level=info msg="Loading containers: done." Jan 24 11:38:35.260142 dockerd[1876]: time="2026-01-24T11:38:35.258318571Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 24 11:38:35.260142 dockerd[1876]: time="2026-01-24T11:38:35.258967631Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Jan 24 11:38:35.260142 dockerd[1876]: time="2026-01-24T11:38:35.259176540Z" level=info msg="Initializing buildkit" Jan 24 11:38:35.651010 dockerd[1876]: time="2026-01-24T11:38:35.648773889Z" level=info msg="Completed buildkit initialization" Jan 24 11:38:35.670309 dockerd[1876]: time="2026-01-24T11:38:35.670113940Z" level=info msg="Daemon has completed initialization" Jan 24 11:38:35.675363 dockerd[1876]: time="2026-01-24T11:38:35.670932238Z" level=info msg="API listen on /run/docker.sock" Jan 24 11:38:35.676328 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 24 11:38:35.676000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:41.906081 containerd[1625]: time="2026-01-24T11:38:41.892034483Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.3\"" Jan 24 11:38:42.433947 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 24 11:38:42.441756 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:38:44.852000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:44.852621 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:38:44.909171 kernel: kauditd_printk_skb: 113 callbacks suppressed Jan 24 11:38:44.913731 kernel: audit: type=1130 audit(1769254724.852:278): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:44.954337 (kubelet)[2124]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:38:45.370866 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2803108595.mount: Deactivated successfully. Jan 24 11:38:45.686836 kubelet[2124]: E0124 11:38:45.649689 2124 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:38:45.837127 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:38:45.932161 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:38:45.940000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:38:45.941947 systemd[1]: kubelet.service: Consumed 1.668s CPU time, 109M memory peak. Jan 24 11:38:45.964346 kernel: audit: type=1131 audit(1769254725.940:279): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:38:55.941910 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Jan 24 11:38:55.971092 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:38:58.026629 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:38:58.025000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:58.053302 kernel: audit: type=1130 audit(1769254738.025:280): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:38:58.110985 (kubelet)[2198]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:38:59.153442 kubelet[2198]: E0124 11:38:59.152504 2198 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:38:59.217742 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:38:59.218650 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:38:59.287000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:38:59.329368 systemd[1]: kubelet.service: Consumed 2.155s CPU time, 108.5M memory peak. Jan 24 11:38:59.346361 kernel: audit: type=1131 audit(1769254739.287:281): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:39:03.335443 containerd[1625]: time="2026-01-24T11:39:03.333455082Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.34.3: active requests=0, bytes read=26484731" Jan 24 11:39:03.364165 containerd[1625]: time="2026-01-24T11:39:03.336871661Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:03.369680 containerd[1625]: time="2026-01-24T11:39:03.365018939Z" level=info msg="ImageCreate event name:\"sha256:aa27095f5619377172f3d59289ccb2ba567ebea93a736d1705be068b2c030b0c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:03.458343 containerd[1625]: time="2026-01-24T11:39:03.451067816Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:5af1030676ceca025742ef5e73a504d11b59be0e5551cdb8c9cf0d3c1231b460\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:03.527369 containerd[1625]: time="2026-01-24T11:39:03.523035779Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.34.3\" with image id \"sha256:aa27095f5619377172f3d59289ccb2ba567ebea93a736d1705be068b2c030b0c\", repo tag \"registry.k8s.io/kube-apiserver:v1.34.3\", repo digest \"registry.k8s.io/kube-apiserver@sha256:5af1030676ceca025742ef5e73a504d11b59be0e5551cdb8c9cf0d3c1231b460\", size \"27064672\" in 21.630109673s" Jan 24 11:39:03.527369 containerd[1625]: time="2026-01-24T11:39:03.524907651Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.3\" returns image reference \"sha256:aa27095f5619377172f3d59289ccb2ba567ebea93a736d1705be068b2c030b0c\"" Jan 24 11:39:03.571806 containerd[1625]: time="2026-01-24T11:39:03.569148878Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.3\"" Jan 24 11:39:09.436801 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Jan 24 11:39:09.451626 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:39:10.872419 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:39:10.871000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:39:10.926499 kernel: audit: type=1130 audit(1769254750.871:282): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:39:10.974924 (kubelet)[2219]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:39:12.058349 kubelet[2219]: E0124 11:39:12.057407 2219 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:39:12.114167 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:39:12.138567 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:39:12.152000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:39:12.163471 systemd[1]: kubelet.service: Consumed 1.500s CPU time, 108.9M memory peak. Jan 24 11:39:12.202666 kernel: audit: type=1131 audit(1769254752.152:283): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:39:12.905641 containerd[1625]: time="2026-01-24T11:39:12.901668896Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:12.905641 containerd[1625]: time="2026-01-24T11:39:12.905469229Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.34.3: active requests=0, bytes read=21154285" Jan 24 11:39:12.912626 containerd[1625]: time="2026-01-24T11:39:12.909672175Z" level=info msg="ImageCreate event name:\"sha256:5826b25d990d7d314d236c8d128f43e443583891f5cdffa7bf8bca50ae9e0942\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:12.927303 containerd[1625]: time="2026-01-24T11:39:12.927037527Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:716a210d31ee5e27053ea0e1a3a3deb4910791a85ba4b1120410b5a4cbcf1954\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:12.941425 containerd[1625]: time="2026-01-24T11:39:12.939634143Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.34.3\" with image id \"sha256:5826b25d990d7d314d236c8d128f43e443583891f5cdffa7bf8bca50ae9e0942\", repo tag \"registry.k8s.io/kube-controller-manager:v1.34.3\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:716a210d31ee5e27053ea0e1a3a3deb4910791a85ba4b1120410b5a4cbcf1954\", size \"22819474\" in 9.369562388s" Jan 24 11:39:12.941425 containerd[1625]: time="2026-01-24T11:39:12.941354192Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.3\" returns image reference \"sha256:5826b25d990d7d314d236c8d128f43e443583891f5cdffa7bf8bca50ae9e0942\"" Jan 24 11:39:12.957590 containerd[1625]: time="2026-01-24T11:39:12.955938922Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.3\"" Jan 24 11:39:22.191693 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Jan 24 11:39:22.219758 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:39:23.781595 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:39:23.781000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:39:23.840364 kernel: audit: type=1130 audit(1769254763.781:284): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:39:23.874999 (kubelet)[2240]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:39:23.954648 containerd[1625]: time="2026-01-24T11:39:23.947983507Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:23.954648 containerd[1625]: time="2026-01-24T11:39:23.951532283Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.34.3: active requests=0, bytes read=15717792" Jan 24 11:39:23.961315 containerd[1625]: time="2026-01-24T11:39:23.961032495Z" level=info msg="ImageCreate event name:\"sha256:aec12dadf56dd45659a682b94571f115a1be02ee4a262b3b5176394f5c030c78\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:23.973507 containerd[1625]: time="2026-01-24T11:39:23.970759398Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:f9a9bc7948fd804ef02255fe82ac2e85d2a66534bae2fe1348c14849260a1fe2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:23.979058 containerd[1625]: time="2026-01-24T11:39:23.978700238Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.34.3\" with image id \"sha256:aec12dadf56dd45659a682b94571f115a1be02ee4a262b3b5176394f5c030c78\", repo tag \"registry.k8s.io/kube-scheduler:v1.34.3\", repo digest \"registry.k8s.io/kube-scheduler@sha256:f9a9bc7948fd804ef02255fe82ac2e85d2a66534bae2fe1348c14849260a1fe2\", size \"17382979\" in 11.022714699s" Jan 24 11:39:23.979058 containerd[1625]: time="2026-01-24T11:39:23.978964051Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.3\" returns image reference \"sha256:aec12dadf56dd45659a682b94571f115a1be02ee4a262b3b5176394f5c030c78\"" Jan 24 11:39:24.027402 containerd[1625]: time="2026-01-24T11:39:24.026449724Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.3\"" Jan 24 11:39:26.992436 kubelet[2240]: E0124 11:39:26.991533 2240 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:39:27.062960 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:39:27.063651 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:39:27.069000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:39:27.071385 systemd[1]: kubelet.service: Consumed 2.602s CPU time, 110.1M memory peak. Jan 24 11:39:27.143575 kernel: audit: type=1131 audit(1769254767.069:285): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:39:35.670083 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1357583015.mount: Deactivated successfully. Jan 24 11:39:37.724721 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Jan 24 11:39:37.887728 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:39:40.584000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:39:40.584867 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:39:40.615399 kernel: audit: type=1130 audit(1769254780.584:286): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:39:40.627638 (kubelet)[2264]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:39:41.434508 kubelet[2264]: E0124 11:39:41.430487 2264 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:39:41.444753 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:39:41.447521 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:39:41.449458 systemd[1]: kubelet.service: Consumed 1.702s CPU time, 109M memory peak. Jan 24 11:39:41.479496 kernel: audit: type=1131 audit(1769254781.448:287): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:39:41.448000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:39:43.382732 containerd[1625]: time="2026-01-24T11:39:43.373718451Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:43.382732 containerd[1625]: time="2026-01-24T11:39:43.380771628Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.34.3: active requests=0, bytes read=25962930" Jan 24 11:39:43.391652 containerd[1625]: time="2026-01-24T11:39:43.390481395Z" level=info msg="ImageCreate event name:\"sha256:36eef8e07bdd6abdc2bbf44041e49480fe499a3cedb0ae054b50daa1a35cf691\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:43.422622 containerd[1625]: time="2026-01-24T11:39:43.422561061Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:7298ab89a103523d02ff4f49bedf9359710af61df92efdc07bac873064f03ed6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:39:43.425700 containerd[1625]: time="2026-01-24T11:39:43.425628882Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.34.3\" with image id \"sha256:36eef8e07bdd6abdc2bbf44041e49480fe499a3cedb0ae054b50daa1a35cf691\", repo tag \"registry.k8s.io/kube-proxy:v1.34.3\", repo digest \"registry.k8s.io/kube-proxy@sha256:7298ab89a103523d02ff4f49bedf9359710af61df92efdc07bac873064f03ed6\", size \"25964312\" in 19.396010733s" Jan 24 11:39:43.426524 containerd[1625]: time="2026-01-24T11:39:43.425923283Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.3\" returns image reference \"sha256:36eef8e07bdd6abdc2bbf44041e49480fe499a3cedb0ae054b50daa1a35cf691\"" Jan 24 11:39:43.446105 containerd[1625]: time="2026-01-24T11:39:43.445946462Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\"" Jan 24 11:39:51.067467 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1411546742.mount: Deactivated successfully. Jan 24 11:39:51.723902 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Jan 24 11:39:51.861826 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:39:53.465793 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:39:53.465000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:39:53.527461 kernel: audit: type=1130 audit(1769254793.465:288): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:39:53.627386 (kubelet)[2293]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:39:54.932942 kubelet[2293]: E0124 11:39:54.931754 2293 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:39:54.943703 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:39:54.944424 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:39:54.945000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:39:54.946818 systemd[1]: kubelet.service: Consumed 2.031s CPU time, 110.6M memory peak. Jan 24 11:39:54.984615 kernel: audit: type=1131 audit(1769254794.945:289): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:40:05.355955 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Jan 24 11:40:05.465095 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:40:08.833575 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:40:08.875000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:40:08.930789 kernel: audit: type=1130 audit(1769254808.875:290): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:40:08.951872 (kubelet)[2350]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:40:09.877959 kubelet[2350]: E0124 11:40:09.835012 2350 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:40:09.937916 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:40:09.940878 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:40:09.954000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:40:09.960733 systemd[1]: kubelet.service: Consumed 2.017s CPU time, 110.7M memory peak. Jan 24 11:40:10.023679 kernel: audit: type=1131 audit(1769254809.954:291): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:40:11.029971 containerd[1625]: time="2026-01-24T11:40:11.028843671Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:40:11.053798 containerd[1625]: time="2026-01-24T11:40:11.041822680Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.1: active requests=0, bytes read=22381861" Jan 24 11:40:11.075896 containerd[1625]: time="2026-01-24T11:40:11.074541342Z" level=info msg="ImageCreate event name:\"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:40:11.134791 containerd[1625]: time="2026-01-24T11:40:11.127418407Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:40:11.154906 containerd[1625]: time="2026-01-24T11:40:11.148937995Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.1\" with image id \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\", size \"22384805\" in 27.702184234s" Jan 24 11:40:11.154906 containerd[1625]: time="2026-01-24T11:40:11.152996067Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\" returns image reference \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\"" Jan 24 11:40:11.187959 containerd[1625]: time="2026-01-24T11:40:11.187392468Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Jan 24 11:40:22.064100 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Jan 24 11:40:22.663691 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:40:24.706778 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount458730917.mount: Deactivated successfully. Jan 24 11:40:24.872633 containerd[1625]: time="2026-01-24T11:40:24.867582888Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:40:24.935792 containerd[1625]: time="2026-01-24T11:40:24.933959194Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=2406" Jan 24 11:40:24.953877 containerd[1625]: time="2026-01-24T11:40:24.953631689Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:40:24.969376 containerd[1625]: time="2026-01-24T11:40:24.967612071Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:40:24.971364 containerd[1625]: time="2026-01-24T11:40:24.970986863Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 13.759294864s" Jan 24 11:40:24.975414 containerd[1625]: time="2026-01-24T11:40:24.973131127Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\"" Jan 24 11:40:25.123148 containerd[1625]: time="2026-01-24T11:40:25.122477011Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.4-0\"" Jan 24 11:40:27.983601 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount493084857.mount: Deactivated successfully. Jan 24 11:40:29.090000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:40:29.088017 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:40:29.145751 kernel: audit: type=1130 audit(1769254829.090:292): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:40:29.664713 (kubelet)[2383]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:40:31.746926 kubelet[2383]: E0124 11:40:31.745619 2383 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:40:31.759413 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:40:31.761614 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:40:31.764000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:40:31.765819 systemd[1]: kubelet.service: Consumed 3.560s CPU time, 110.2M memory peak. Jan 24 11:40:31.821964 kernel: audit: type=1131 audit(1769254831.764:293): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:40:41.978863 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. Jan 24 11:40:42.024697 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:40:44.324569 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:40:44.323000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:40:44.348355 kernel: audit: type=1130 audit(1769254844.323:294): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:40:44.387127 (kubelet)[2440]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:40:45.100593 kubelet[2440]: E0124 11:40:45.098866 2440 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:40:45.133783 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:40:45.134370 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:40:45.137000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:40:45.139503 systemd[1]: kubelet.service: Consumed 1.331s CPU time, 110.3M memory peak. Jan 24 11:40:45.170356 kernel: audit: type=1131 audit(1769254845.137:295): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:40:55.342869 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. Jan 24 11:40:55.486454 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:41:01.458079 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:41:01.461000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:41:01.529579 kernel: audit: type=1130 audit(1769254861.461:296): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:41:01.543965 (kubelet)[2457]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:41:02.979475 kubelet[2457]: E0124 11:41:02.977002 2457 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:41:02.997713 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:41:03.008000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:41:02.999867 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:41:03.009458 systemd[1]: kubelet.service: Consumed 2.825s CPU time, 110.6M memory peak. Jan 24 11:41:03.044464 kernel: audit: type=1131 audit(1769254863.008:297): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:41:09.142961 containerd[1625]: time="2026-01-24T11:41:09.142413470Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.4-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:41:09.150045 containerd[1625]: time="2026-01-24T11:41:09.146988958Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.4-0: active requests=0, bytes read=73002391" Jan 24 11:41:09.153334 containerd[1625]: time="2026-01-24T11:41:09.153111710Z" level=info msg="ImageCreate event name:\"sha256:5f1f5298c888daa46c4409ff4cefe5ca9d16e479419f94cdb5f5d5563dac0115\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:41:09.196935 containerd[1625]: time="2026-01-24T11:41:09.186655619Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:e36c081683425b5b3bc1425bc508b37e7107bb65dfa9367bf5a80125d431fa19\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:41:09.219979 containerd[1625]: time="2026-01-24T11:41:09.207942110Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.4-0\" with image id \"sha256:5f1f5298c888daa46c4409ff4cefe5ca9d16e479419f94cdb5f5d5563dac0115\", repo tag \"registry.k8s.io/etcd:3.6.4-0\", repo digest \"registry.k8s.io/etcd@sha256:e36c081683425b5b3bc1425bc508b37e7107bb65dfa9367bf5a80125d431fa19\", size \"74311308\" in 44.074817503s" Jan 24 11:41:09.232415 containerd[1625]: time="2026-01-24T11:41:09.232142710Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.4-0\" returns image reference \"sha256:5f1f5298c888daa46c4409ff4cefe5ca9d16e479419f94cdb5f5d5563dac0115\"" Jan 24 11:41:13.246540 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 12. Jan 24 11:41:13.269721 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:41:15.540337 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:41:15.540000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:41:15.568730 kernel: audit: type=1130 audit(1769254875.540:298): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:41:15.672118 (kubelet)[2500]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:41:17.128551 kubelet[2500]: E0124 11:41:17.127050 2500 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:41:17.138100 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:41:17.138907 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:41:17.140587 systemd[1]: kubelet.service: Consumed 1.469s CPU time, 110.8M memory peak. Jan 24 11:41:17.138000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:41:17.165510 kernel: audit: type=1131 audit(1769254877.138:299): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:41:24.980000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:41:24.984759 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:41:24.985009 systemd[1]: kubelet.service: Consumed 1.469s CPU time, 110.8M memory peak. Jan 24 11:41:24.997678 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:41:24.980000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:41:25.136575 kernel: audit: type=1130 audit(1769254884.980:300): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:41:25.136852 kernel: audit: type=1131 audit(1769254884.980:301): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:41:25.196843 systemd[1]: Reload requested from client PID 2514 ('systemctl') (unit session-10.scope)... Jan 24 11:41:25.199125 systemd[1]: Reloading... Jan 24 11:41:25.828481 zram_generator::config[2558]: No configuration found. Jan 24 11:41:27.866390 systemd[1]: Reloading finished in 2658 ms. Jan 24 11:41:28.023000 audit: BPF prog-id=61 op=LOAD Jan 24 11:41:28.023000 audit: BPF prog-id=41 op=UNLOAD Jan 24 11:41:28.057804 kernel: audit: type=1334 audit(1769254888.023:302): prog-id=61 op=LOAD Jan 24 11:41:28.057989 kernel: audit: type=1334 audit(1769254888.023:303): prog-id=41 op=UNLOAD Jan 24 11:41:28.058128 kernel: audit: type=1334 audit(1769254888.024:304): prog-id=62 op=LOAD Jan 24 11:41:28.058172 kernel: audit: type=1334 audit(1769254888.024:305): prog-id=63 op=LOAD Jan 24 11:41:28.058546 kernel: audit: type=1334 audit(1769254888.024:306): prog-id=54 op=UNLOAD Jan 24 11:41:28.058820 kernel: audit: type=1334 audit(1769254888.024:307): prog-id=55 op=UNLOAD Jan 24 11:41:28.058873 kernel: audit: type=1334 audit(1769254888.031:308): prog-id=64 op=LOAD Jan 24 11:41:28.059108 kernel: audit: type=1334 audit(1769254888.031:309): prog-id=51 op=UNLOAD Jan 24 11:41:28.024000 audit: BPF prog-id=62 op=LOAD Jan 24 11:41:28.024000 audit: BPF prog-id=63 op=LOAD Jan 24 11:41:28.024000 audit: BPF prog-id=54 op=UNLOAD Jan 24 11:41:28.024000 audit: BPF prog-id=55 op=UNLOAD Jan 24 11:41:28.031000 audit: BPF prog-id=64 op=LOAD Jan 24 11:41:28.031000 audit: BPF prog-id=51 op=UNLOAD Jan 24 11:41:28.031000 audit: BPF prog-id=65 op=LOAD Jan 24 11:41:28.031000 audit: BPF prog-id=66 op=LOAD Jan 24 11:41:28.031000 audit: BPF prog-id=52 op=UNLOAD Jan 24 11:41:28.031000 audit: BPF prog-id=53 op=UNLOAD Jan 24 11:41:28.038000 audit: BPF prog-id=67 op=LOAD Jan 24 11:41:28.038000 audit: BPF prog-id=57 op=UNLOAD Jan 24 11:41:28.044000 audit: BPF prog-id=68 op=LOAD Jan 24 11:41:28.044000 audit: BPF prog-id=42 op=UNLOAD Jan 24 11:41:28.044000 audit: BPF prog-id=69 op=LOAD Jan 24 11:41:28.044000 audit: BPF prog-id=70 op=LOAD Jan 24 11:41:28.044000 audit: BPF prog-id=43 op=UNLOAD Jan 24 11:41:28.044000 audit: BPF prog-id=44 op=UNLOAD Jan 24 11:41:28.048000 audit: BPF prog-id=71 op=LOAD Jan 24 11:41:28.048000 audit: BPF prog-id=48 op=UNLOAD Jan 24 11:41:28.049000 audit: BPF prog-id=72 op=LOAD Jan 24 11:41:28.049000 audit: BPF prog-id=73 op=LOAD Jan 24 11:41:28.049000 audit: BPF prog-id=49 op=UNLOAD Jan 24 11:41:28.049000 audit: BPF prog-id=50 op=UNLOAD Jan 24 11:41:28.054000 audit: BPF prog-id=74 op=LOAD Jan 24 11:41:28.054000 audit: BPF prog-id=45 op=UNLOAD Jan 24 11:41:28.054000 audit: BPF prog-id=75 op=LOAD Jan 24 11:41:28.054000 audit: BPF prog-id=76 op=LOAD Jan 24 11:41:28.054000 audit: BPF prog-id=46 op=UNLOAD Jan 24 11:41:28.054000 audit: BPF prog-id=47 op=UNLOAD Jan 24 11:41:28.059000 audit: BPF prog-id=77 op=LOAD Jan 24 11:41:28.059000 audit: BPF prog-id=56 op=UNLOAD Jan 24 11:41:28.068000 audit: BPF prog-id=78 op=LOAD Jan 24 11:41:28.068000 audit: BPF prog-id=58 op=UNLOAD Jan 24 11:41:28.068000 audit: BPF prog-id=79 op=LOAD Jan 24 11:41:28.068000 audit: BPF prog-id=80 op=LOAD Jan 24 11:41:28.068000 audit: BPF prog-id=59 op=UNLOAD Jan 24 11:41:28.068000 audit: BPF prog-id=60 op=UNLOAD Jan 24 11:41:28.183000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:41:28.183876 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 24 11:41:28.184030 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 24 11:41:28.184988 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:41:28.185063 systemd[1]: kubelet.service: Consumed 449ms CPU time, 98.4M memory peak. Jan 24 11:41:28.214129 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:41:29.628000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:41:29.629922 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:41:29.681612 (kubelet)[2609]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 24 11:41:31.006157 kubelet[2609]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 24 11:41:31.006157 kubelet[2609]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 24 11:41:31.030125 kubelet[2609]: I0124 11:41:31.025963 2609 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 24 11:41:31.414815 kubelet[2609]: I0124 11:41:31.410600 2609 server.go:529] "Kubelet version" kubeletVersion="v1.34.1" Jan 24 11:41:31.414815 kubelet[2609]: I0124 11:41:31.412475 2609 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 24 11:41:31.414815 kubelet[2609]: I0124 11:41:31.412914 2609 watchdog_linux.go:95] "Systemd watchdog is not enabled" Jan 24 11:41:31.414815 kubelet[2609]: I0124 11:41:31.412938 2609 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 24 11:41:31.420110 kubelet[2609]: I0124 11:41:31.416072 2609 server.go:956] "Client rotation is on, will bootstrap in background" Jan 24 11:41:31.566785 kubelet[2609]: E0124 11:41:31.559039 2609 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.26:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 24 11:41:31.591288 kubelet[2609]: I0124 11:41:31.590901 2609 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 24 11:41:31.817558 kubelet[2609]: I0124 11:41:31.816160 2609 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 24 11:41:32.056584 kubelet[2609]: I0124 11:41:32.050896 2609 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Jan 24 11:41:32.098869 kubelet[2609]: I0124 11:41:32.064870 2609 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 24 11:41:32.098869 kubelet[2609]: I0124 11:41:32.064986 2609 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 24 11:41:32.098869 kubelet[2609]: I0124 11:41:32.090167 2609 topology_manager.go:138] "Creating topology manager with none policy" Jan 24 11:41:32.098869 kubelet[2609]: I0124 11:41:32.090529 2609 container_manager_linux.go:306] "Creating device plugin manager" Jan 24 11:41:32.135823 kubelet[2609]: I0124 11:41:32.091662 2609 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Jan 24 11:41:32.161729 kubelet[2609]: I0124 11:41:32.160660 2609 state_mem.go:36] "Initialized new in-memory state store" Jan 24 11:41:32.189149 kubelet[2609]: I0124 11:41:32.187592 2609 kubelet.go:475] "Attempting to sync node with API server" Jan 24 11:41:32.189149 kubelet[2609]: I0124 11:41:32.188329 2609 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 24 11:41:32.244105 kubelet[2609]: I0124 11:41:32.189165 2609 kubelet.go:387] "Adding apiserver pod source" Jan 24 11:41:32.244105 kubelet[2609]: I0124 11:41:32.192833 2609 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 24 11:41:32.246860 kubelet[2609]: E0124 11:41:32.246145 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.26:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 24 11:41:32.259803 kubelet[2609]: E0124 11:41:32.247629 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.26:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 24 11:41:32.348713 kubelet[2609]: I0124 11:41:32.343483 2609 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 24 11:41:32.375673 kubelet[2609]: I0124 11:41:32.371516 2609 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 24 11:41:32.375673 kubelet[2609]: I0124 11:41:32.374553 2609 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Jan 24 11:41:32.381807 kubelet[2609]: W0124 11:41:32.379487 2609 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 24 11:41:32.468076 kubelet[2609]: I0124 11:41:32.467869 2609 server.go:1262] "Started kubelet" Jan 24 11:41:32.478462 kubelet[2609]: I0124 11:41:32.477405 2609 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 24 11:41:32.537414 kubelet[2609]: I0124 11:41:32.531442 2609 server.go:310] "Adding debug handlers to kubelet server" Jan 24 11:41:32.537414 kubelet[2609]: I0124 11:41:32.532553 2609 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 24 11:41:32.537414 kubelet[2609]: I0124 11:41:32.532803 2609 server_v1.go:49] "podresources" method="list" useActivePods=true Jan 24 11:41:32.537414 kubelet[2609]: I0124 11:41:32.537150 2609 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 24 11:41:32.561580 kubelet[2609]: I0124 11:41:32.561531 2609 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 24 11:41:32.579978 kubelet[2609]: E0124 11:41:32.564493 2609 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.26:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.26:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.188da7f366a4ec6b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-24 11:41:32.467448939 +0000 UTC m=+2.736922269,LastTimestamp:2026-01-24 11:41:32.467448939 +0000 UTC m=+2.736922269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 24 11:41:32.651419 kubelet[2609]: I0124 11:41:32.650388 2609 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 24 11:41:32.672044 kubelet[2609]: I0124 11:41:32.671942 2609 volume_manager.go:313] "Starting Kubelet Volume Manager" Jan 24 11:41:32.683483 kubelet[2609]: E0124 11:41:32.680908 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:32.684793 kubelet[2609]: I0124 11:41:32.684764 2609 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 24 11:41:32.713808 kubelet[2609]: E0124 11:41:32.695916 2609 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.26:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.26:6443: connect: connection refused" interval="200ms" Jan 24 11:41:32.715754 kubelet[2609]: I0124 11:41:32.715586 2609 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 24 11:41:32.736506 kubelet[2609]: E0124 11:41:32.735690 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.26:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 24 11:41:32.742697 kubelet[2609]: I0124 11:41:32.742621 2609 reconciler.go:29] "Reconciler: start to sync state" Jan 24 11:41:32.763757 kubelet[2609]: E0124 11:41:32.763707 2609 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 24 11:41:32.772413 kubelet[2609]: I0124 11:41:32.772351 2609 factory.go:223] Registration of the containerd container factory successfully Jan 24 11:41:32.772413 kubelet[2609]: I0124 11:41:32.772392 2609 factory.go:223] Registration of the systemd container factory successfully Jan 24 11:41:32.783850 kubelet[2609]: E0124 11:41:32.783437 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:32.885643 kubelet[2609]: E0124 11:41:32.884909 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:32.921760 kubelet[2609]: E0124 11:41:32.919912 2609 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.26:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.26:6443: connect: connection refused" interval="400ms" Jan 24 11:41:32.927000 audit[2631]: NETFILTER_CFG table=mangle:42 family=2 entries=2 op=nft_register_chain pid=2631 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:32.946742 kernel: kauditd_printk_skb: 34 callbacks suppressed Jan 24 11:41:32.966389 kernel: audit: type=1325 audit(1769254892.927:344): table=mangle:42 family=2 entries=2 op=nft_register_chain pid=2631 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:32.927000 audit[2631]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffd41f8f400 a2=0 a3=0 items=0 ppid=2609 pid=2631 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:32.991959 kubelet[2609]: E0124 11:41:32.991091 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.030132 kernel: audit: type=1300 audit(1769254892.927:344): arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffd41f8f400 a2=0 a3=0 items=0 ppid=2609 pid=2631 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.054824 kernel: audit: type=1327 audit(1769254892.927:344): proctitle=69707461626C6573002D770035002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 24 11:41:33.055802 kernel: audit: type=1325 audit(1769254892.972:345): table=filter:43 family=2 entries=1 op=nft_register_chain pid=2632 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:32.927000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 24 11:41:32.972000 audit[2632]: NETFILTER_CFG table=filter:43 family=2 entries=1 op=nft_register_chain pid=2632 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:32.972000 audit[2632]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc51999a10 a2=0 a3=0 items=0 ppid=2609 pid=2632 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:32.972000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 24 11:41:33.149807 kernel: audit: type=1300 audit(1769254892.972:345): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc51999a10 a2=0 a3=0 items=0 ppid=2609 pid=2632 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.156373 kernel: audit: type=1327 audit(1769254892.972:345): proctitle=69707461626C6573002D770035002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 24 11:41:33.162100 kubelet[2609]: E0124 11:41:33.133890 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.162000 audit[2634]: NETFILTER_CFG table=filter:44 family=2 entries=2 op=nft_register_chain pid=2634 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:33.231638 kernel: audit: type=1325 audit(1769254893.162:346): table=filter:44 family=2 entries=2 op=nft_register_chain pid=2634 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:33.235354 kernel: audit: type=1300 audit(1769254893.162:346): arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffdef239ed0 a2=0 a3=0 items=0 ppid=2609 pid=2634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.162000 audit[2634]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffdef239ed0 a2=0 a3=0 items=0 ppid=2609 pid=2634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.244427 kubelet[2609]: E0124 11:41:33.243781 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.291560 kernel: audit: type=1327 audit(1769254893.162:346): proctitle=69707461626C6573002D770035002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 24 11:41:33.162000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 24 11:41:33.344061 kubelet[2609]: E0124 11:41:33.343686 2609 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.26:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.26:6443: connect: connection refused" interval="800ms" Jan 24 11:41:33.347494 kubelet[2609]: E0124 11:41:33.344610 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.356000 audit[2638]: NETFILTER_CFG table=filter:45 family=2 entries=2 op=nft_register_chain pid=2638 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:33.400877 kubelet[2609]: I0124 11:41:33.400668 2609 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 24 11:41:33.404485 kubelet[2609]: I0124 11:41:33.402619 2609 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 24 11:41:33.404485 kubelet[2609]: I0124 11:41:33.403004 2609 state_mem.go:36] "Initialized new in-memory state store" Jan 24 11:41:33.436852 kernel: audit: type=1325 audit(1769254893.356:347): table=filter:45 family=2 entries=2 op=nft_register_chain pid=2638 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:33.356000 audit[2638]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffd2e2cf1c0 a2=0 a3=0 items=0 ppid=2609 pid=2638 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.356000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 24 11:41:33.451812 kubelet[2609]: E0124 11:41:33.450504 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.466366 kubelet[2609]: I0124 11:41:33.464781 2609 policy_none.go:49] "None policy: Start" Jan 24 11:41:33.470488 kubelet[2609]: I0124 11:41:33.468795 2609 memory_manager.go:187] "Starting memorymanager" policy="None" Jan 24 11:41:33.470488 kubelet[2609]: I0124 11:41:33.468975 2609 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Jan 24 11:41:33.487655 kubelet[2609]: I0124 11:41:33.484865 2609 policy_none.go:47] "Start" Jan 24 11:41:33.544000 audit[2641]: NETFILTER_CFG table=filter:46 family=2 entries=1 op=nft_register_rule pid=2641 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:33.544000 audit[2641]: SYSCALL arch=c000003e syscall=46 success=yes exit=924 a0=3 a1=7fffea0be750 a2=0 a3=0 items=0 ppid=2609 pid=2641 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.544000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F380000002D2D737263003132372E Jan 24 11:41:33.554667 kubelet[2609]: E0124 11:41:33.554021 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.555011 kubelet[2609]: I0124 11:41:33.554880 2609 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Jan 24 11:41:33.565000 audit[2642]: NETFILTER_CFG table=mangle:47 family=2 entries=1 op=nft_register_chain pid=2642 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:33.565000 audit[2642]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd252c08d0 a2=0 a3=0 items=0 ppid=2609 pid=2642 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.565000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 24 11:41:33.586000 audit[2643]: NETFILTER_CFG table=mangle:48 family=10 entries=2 op=nft_register_chain pid=2643 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:41:33.586000 audit[2643]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffe419b8d90 a2=0 a3=0 items=0 ppid=2609 pid=2643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.586000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 24 11:41:33.593000 audit[2645]: NETFILTER_CFG table=nat:49 family=2 entries=1 op=nft_register_chain pid=2645 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:33.593000 audit[2645]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe9a3d49b0 a2=0 a3=0 items=0 ppid=2609 pid=2645 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.593000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 24 11:41:33.648474 kubelet[2609]: I0124 11:41:33.637803 2609 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Jan 24 11:41:33.648474 kubelet[2609]: I0124 11:41:33.643095 2609 status_manager.go:244] "Starting to sync pod status with apiserver" Jan 24 11:41:33.648474 kubelet[2609]: I0124 11:41:33.644474 2609 kubelet.go:2427] "Starting kubelet main sync loop" Jan 24 11:41:33.648474 kubelet[2609]: E0124 11:41:33.644805 2609 kubelet.go:2451] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 24 11:41:33.652970 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 24 11:41:33.658086 kubelet[2609]: E0124 11:41:33.655627 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.658086 kubelet[2609]: E0124 11:41:33.656009 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.26:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 24 11:41:33.682485 kubelet[2609]: E0124 11:41:33.681981 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.26:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 24 11:41:33.693000 audit[2646]: NETFILTER_CFG table=filter:50 family=2 entries=1 op=nft_register_chain pid=2646 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:41:33.693000 audit[2646]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd6d1677c0 a2=0 a3=0 items=0 ppid=2609 pid=2646 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.693000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 24 11:41:33.703000 audit[2647]: NETFILTER_CFG table=mangle:51 family=10 entries=1 op=nft_register_chain pid=2647 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:41:33.703000 audit[2647]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc12bd11f0 a2=0 a3=0 items=0 ppid=2609 pid=2647 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.703000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 24 11:41:33.734000 audit[2648]: NETFILTER_CFG table=nat:52 family=10 entries=1 op=nft_register_chain pid=2648 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:41:33.734000 audit[2648]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd46198a80 a2=0 a3=0 items=0 ppid=2609 pid=2648 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.734000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 24 11:41:33.743688 kubelet[2609]: E0124 11:41:33.743520 2609 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.26:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 24 11:41:33.746084 kubelet[2609]: E0124 11:41:33.746052 2609 kubelet.go:2451] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 24 11:41:33.746767 kubelet[2609]: E0124 11:41:33.746736 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.26:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 24 11:41:33.741000 audit[2649]: NETFILTER_CFG table=filter:53 family=10 entries=1 op=nft_register_chain pid=2649 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:41:33.741000 audit[2649]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe8961c6a0 a2=0 a3=0 items=0 ppid=2609 pid=2649 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:33.741000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 24 11:41:33.747750 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 24 11:41:33.755843 kubelet[2609]: E0124 11:41:33.755816 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.857639 kubelet[2609]: E0124 11:41:33.857560 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.883137 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 24 11:41:33.947572 kubelet[2609]: E0124 11:41:33.947113 2609 kubelet.go:2451] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 24 11:41:33.963609 kubelet[2609]: E0124 11:41:33.963034 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:41:33.979458 kubelet[2609]: E0124 11:41:33.975988 2609 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 24 11:41:33.979458 kubelet[2609]: I0124 11:41:33.977085 2609 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 24 11:41:33.989772 kubelet[2609]: I0124 11:41:33.981588 2609 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 24 11:41:34.021547 kubelet[2609]: I0124 11:41:34.018013 2609 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 24 11:41:34.030641 kubelet[2609]: E0124 11:41:34.030454 2609 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 24 11:41:34.031121 kubelet[2609]: E0124 11:41:34.030941 2609 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 24 11:41:34.040670 kubelet[2609]: E0124 11:41:33.832805 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.26:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 24 11:41:34.122376 kubelet[2609]: I0124 11:41:34.119731 2609 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:41:34.122849 kubelet[2609]: E0124 11:41:34.122720 2609 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.26:6443/api/v1/nodes\": dial tcp 10.0.0.26:6443: connect: connection refused" node="localhost" Jan 24 11:41:34.151099 kubelet[2609]: E0124 11:41:34.150492 2609 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.26:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.26:6443: connect: connection refused" interval="1.6s" Jan 24 11:41:34.431915 kubelet[2609]: I0124 11:41:34.420810 2609 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:41:34.446005 kubelet[2609]: E0124 11:41:34.445953 2609 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.26:6443/api/v1/nodes\": dial tcp 10.0.0.26:6443: connect: connection refused" node="localhost" Jan 24 11:41:34.526618 kubelet[2609]: E0124 11:41:34.524587 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.26:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 24 11:41:34.646021 kubelet[2609]: I0124 11:41:34.639978 2609 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a1c0b96845ba45e5728846e5b1af0a4c-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"a1c0b96845ba45e5728846e5b1af0a4c\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:41:34.654529 kubelet[2609]: I0124 11:41:34.651398 2609 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a1c0b96845ba45e5728846e5b1af0a4c-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"a1c0b96845ba45e5728846e5b1af0a4c\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:41:34.654529 kubelet[2609]: I0124 11:41:34.653551 2609 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:41:34.654529 kubelet[2609]: I0124 11:41:34.653869 2609 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:41:34.654529 kubelet[2609]: I0124 11:41:34.654101 2609 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:41:34.654529 kubelet[2609]: I0124 11:41:34.654365 2609 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:41:34.655526 kubelet[2609]: I0124 11:41:34.654394 2609 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:41:34.655526 kubelet[2609]: I0124 11:41:34.654452 2609 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a1c0b96845ba45e5728846e5b1af0a4c-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"a1c0b96845ba45e5728846e5b1af0a4c\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:41:34.776377 kubelet[2609]: I0124 11:41:34.774714 2609 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/07ca0cbf79ad6ba9473d8e9f7715e571-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"07ca0cbf79ad6ba9473d8e9f7715e571\") " pod="kube-system/kube-scheduler-localhost" Jan 24 11:41:34.782684 systemd[1]: Created slice kubepods-burstable-poda1c0b96845ba45e5728846e5b1af0a4c.slice - libcontainer container kubepods-burstable-poda1c0b96845ba45e5728846e5b1af0a4c.slice. Jan 24 11:41:35.034957 kubelet[2609]: I0124 11:41:35.030676 2609 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:41:35.041031 kubelet[2609]: E0124 11:41:35.034843 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:35.046623 kubelet[2609]: E0124 11:41:35.046573 2609 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.26:6443/api/v1/nodes\": dial tcp 10.0.0.26:6443: connect: connection refused" node="localhost" Jan 24 11:41:35.061760 kubelet[2609]: E0124 11:41:35.058581 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:35.089915 systemd[1]: Created slice kubepods-burstable-pod5bbfee13ce9e07281eca876a0b8067f2.slice - libcontainer container kubepods-burstable-pod5bbfee13ce9e07281eca876a0b8067f2.slice. Jan 24 11:41:35.347625 containerd[1625]: time="2026-01-24T11:41:35.345603061Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:a1c0b96845ba45e5728846e5b1af0a4c,Namespace:kube-system,Attempt:0,}" Jan 24 11:41:35.371603 kubelet[2609]: E0124 11:41:35.367733 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:35.448912 kubelet[2609]: E0124 11:41:35.447501 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:35.485449 kubelet[2609]: E0124 11:41:35.462837 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.26:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 24 11:41:35.486049 containerd[1625]: time="2026-01-24T11:41:35.477441074Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:5bbfee13ce9e07281eca876a0b8067f2,Namespace:kube-system,Attempt:0,}" Jan 24 11:41:35.566065 systemd[1]: Created slice kubepods-burstable-pod07ca0cbf79ad6ba9473d8e9f7715e571.slice - libcontainer container kubepods-burstable-pod07ca0cbf79ad6ba9473d8e9f7715e571.slice. Jan 24 11:41:35.581789 kubelet[2609]: E0124 11:41:35.579071 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:35.641039 kubelet[2609]: E0124 11:41:35.629450 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:35.641613 containerd[1625]: time="2026-01-24T11:41:35.631536604Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:07ca0cbf79ad6ba9473d8e9f7715e571,Namespace:kube-system,Attempt:0,}" Jan 24 11:41:35.765610 kubelet[2609]: E0124 11:41:35.763908 2609 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.26:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.26:6443: connect: connection refused" interval="3.2s" Jan 24 11:41:35.880617 kubelet[2609]: I0124 11:41:35.879904 2609 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:41:35.885669 kubelet[2609]: E0124 11:41:35.885616 2609 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.26:6443/api/v1/nodes\": dial tcp 10.0.0.26:6443: connect: connection refused" node="localhost" Jan 24 11:41:36.458743 kubelet[2609]: E0124 11:41:36.457816 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.26:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 24 11:41:36.828785 kubelet[2609]: E0124 11:41:36.825787 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.26:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 24 11:41:37.092599 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1587225720.mount: Deactivated successfully. Jan 24 11:41:37.550533 containerd[1625]: time="2026-01-24T11:41:37.542764865Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 24 11:41:37.575993 containerd[1625]: time="2026-01-24T11:41:37.575948966Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Jan 24 11:41:37.582579 kubelet[2609]: I0124 11:41:37.579963 2609 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:41:37.582579 kubelet[2609]: E0124 11:41:37.581719 2609 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.26:6443/api/v1/nodes\": dial tcp 10.0.0.26:6443: connect: connection refused" node="localhost" Jan 24 11:41:37.645553 containerd[1625]: time="2026-01-24T11:41:37.643628055Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 24 11:41:37.666624 containerd[1625]: time="2026-01-24T11:41:37.666569165Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 24 11:41:37.671479 containerd[1625]: time="2026-01-24T11:41:37.671412059Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Jan 24 11:41:37.693043 containerd[1625]: time="2026-01-24T11:41:37.684354323Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 24 11:41:37.732859 containerd[1625]: time="2026-01-24T11:41:37.730979079Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 24 11:41:37.734508 kubelet[2609]: E0124 11:41:37.730922 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.26:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 24 11:41:37.743582 containerd[1625]: time="2026-01-24T11:41:37.742047177Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Jan 24 11:41:37.746910 containerd[1625]: time="2026-01-24T11:41:37.745981167Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 2.20197794s" Jan 24 11:41:37.786917 containerd[1625]: time="2026-01-24T11:41:37.783958301Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 2.31099659s" Jan 24 11:41:37.826508 containerd[1625]: time="2026-01-24T11:41:37.825892094Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 2.10120079s" Jan 24 11:41:38.076870 kubelet[2609]: E0124 11:41:38.076342 2609 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.26:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 24 11:41:38.971073 kubelet[2609]: E0124 11:41:38.970724 2609 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.26:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.26:6443: connect: connection refused" interval="6.4s" Jan 24 11:41:39.078961 containerd[1625]: time="2026-01-24T11:41:39.078888466Z" level=info msg="connecting to shim b125664fc07e6448fa2ae800c4d35891c37744064bcd79d98616ad893b99d3fe" address="unix:///run/containerd/s/afd1f8e3e32a97be4a50ecf770aeae76fb4ba5dee76bba171604355827407b49" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:41:39.165687 containerd[1625]: time="2026-01-24T11:41:39.165430825Z" level=info msg="connecting to shim 5e9dcb862ae1b53dbad14b8ca80587e8a60dfc5660e5c91d8bda9d385a7a887e" address="unix:///run/containerd/s/486e5a14bdfba552debb04825c3a379ed542663b09e58d94f194a16faeff4d1d" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:41:39.735674 kubelet[2609]: E0124 11:41:39.698954 2609 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.26:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.26:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.188da7f366a4ec6b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-24 11:41:32.467448939 +0000 UTC m=+2.736922269,LastTimestamp:2026-01-24 11:41:32.467448939 +0000 UTC m=+2.736922269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 24 11:41:39.788041 containerd[1625]: time="2026-01-24T11:41:39.785037220Z" level=info msg="connecting to shim a9444327836eedc937d6e5c604421e4e8b6cf27385034aebf9af40ac694b8db9" address="unix:///run/containerd/s/6953372d851b3c59b64a6d91f6389ab8cd31e8d7231f7e148621e804f1dac107" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:41:40.328045 kubelet[2609]: E0124 11:41:40.327521 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.26:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 24 11:41:41.040177 kubelet[2609]: E0124 11:41:41.033783 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.26:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 24 11:41:41.059834 kubelet[2609]: E0124 11:41:41.042640 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.26:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 24 11:41:41.053644 systemd[1]: Started cri-containerd-5e9dcb862ae1b53dbad14b8ca80587e8a60dfc5660e5c91d8bda9d385a7a887e.scope - libcontainer container 5e9dcb862ae1b53dbad14b8ca80587e8a60dfc5660e5c91d8bda9d385a7a887e. Jan 24 11:41:41.062634 kubelet[2609]: I0124 11:41:41.061938 2609 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:41:41.066356 kubelet[2609]: E0124 11:41:41.065467 2609 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.26:6443/api/v1/nodes\": dial tcp 10.0.0.26:6443: connect: connection refused" node="localhost" Jan 24 11:41:42.132000 audit: BPF prog-id=81 op=LOAD Jan 24 11:41:42.144644 kernel: kauditd_printk_skb: 26 callbacks suppressed Jan 24 11:41:42.144901 kernel: audit: type=1334 audit(1769254902.132:356): prog-id=81 op=LOAD Jan 24 11:41:42.175623 kernel: audit: type=1334 audit(1769254902.144:357): prog-id=82 op=LOAD Jan 24 11:41:42.239167 kernel: audit: type=1300 audit(1769254902.144:357): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00027a238 a2=98 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.239393 kernel: audit: type=1327 audit(1769254902.144:357): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:42.144000 audit: BPF prog-id=82 op=LOAD Jan 24 11:41:42.144000 audit[2698]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00027a238 a2=98 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.144000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:42.144000 audit: BPF prog-id=82 op=UNLOAD Jan 24 11:41:42.341461 kernel: audit: type=1334 audit(1769254902.144:358): prog-id=82 op=UNLOAD Jan 24 11:41:42.341849 kernel: audit: type=1300 audit(1769254902.144:358): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.144000 audit[2698]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.398042 kernel: audit: type=1327 audit(1769254902.144:358): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:42.525060 kernel: audit: type=1334 audit(1769254902.162:359): prog-id=83 op=LOAD Jan 24 11:41:42.525466 kernel: audit: type=1300 audit(1769254902.162:359): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00027a488 a2=98 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.555488 kernel: audit: type=1327 audit(1769254902.162:359): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:42.144000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:42.162000 audit: BPF prog-id=83 op=LOAD Jan 24 11:41:42.162000 audit[2698]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00027a488 a2=98 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.162000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:42.163000 audit: BPF prog-id=84 op=LOAD Jan 24 11:41:42.163000 audit[2698]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00027a218 a2=98 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.163000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:42.163000 audit: BPF prog-id=84 op=UNLOAD Jan 24 11:41:42.163000 audit[2698]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.163000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:42.182000 audit: BPF prog-id=83 op=UNLOAD Jan 24 11:41:42.182000 audit[2698]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.182000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:42.233000 audit: BPF prog-id=85 op=LOAD Jan 24 11:41:42.233000 audit[2698]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00027a6e8 a2=98 a3=0 items=0 ppid=2676 pid=2698 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:42.233000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3565396463623836326165316235336462616431346238636138303538 Jan 24 11:41:43.151711 kubelet[2609]: E0124 11:41:43.143334 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.26:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 24 11:41:43.168371 containerd[1625]: time="2026-01-24T11:41:43.165540729Z" level=error msg="get state for 5e9dcb862ae1b53dbad14b8ca80587e8a60dfc5660e5c91d8bda9d385a7a887e" error="context deadline exceeded" Jan 24 11:41:43.168371 containerd[1625]: time="2026-01-24T11:41:43.165588508Z" level=warning msg="unknown status" status=0 Jan 24 11:41:43.261429 systemd[1]: Started cri-containerd-b125664fc07e6448fa2ae800c4d35891c37744064bcd79d98616ad893b99d3fe.scope - libcontainer container b125664fc07e6448fa2ae800c4d35891c37744064bcd79d98616ad893b99d3fe. Jan 24 11:41:43.343475 systemd[1]: Started cri-containerd-a9444327836eedc937d6e5c604421e4e8b6cf27385034aebf9af40ac694b8db9.scope - libcontainer container a9444327836eedc937d6e5c604421e4e8b6cf27385034aebf9af40ac694b8db9. Jan 24 11:41:44.069461 kubelet[2609]: E0124 11:41:44.063439 2609 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 24 11:41:44.176827 containerd[1625]: time="2026-01-24T11:41:44.176647968Z" level=error msg="ttrpc: received message on inactive stream" stream=3 Jan 24 11:41:44.252000 audit: BPF prog-id=86 op=LOAD Jan 24 11:41:44.275000 audit: BPF prog-id=87 op=LOAD Jan 24 11:41:44.275000 audit[2706]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0000ea238 a2=98 a3=0 items=0 ppid=2664 pid=2706 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.275000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6231323536363466633037653634343866613261653830306334643335 Jan 24 11:41:44.276000 audit: BPF prog-id=87 op=UNLOAD Jan 24 11:41:44.276000 audit[2706]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2664 pid=2706 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.276000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6231323536363466633037653634343866613261653830306334643335 Jan 24 11:41:44.282000 audit: BPF prog-id=88 op=LOAD Jan 24 11:41:44.282000 audit[2706]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0000ea488 a2=98 a3=0 items=0 ppid=2664 pid=2706 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.282000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6231323536363466633037653634343866613261653830306334643335 Jan 24 11:41:44.284000 audit: BPF prog-id=89 op=LOAD Jan 24 11:41:44.284000 audit[2706]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0000ea218 a2=98 a3=0 items=0 ppid=2664 pid=2706 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.284000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6231323536363466633037653634343866613261653830306334643335 Jan 24 11:41:44.286000 audit: BPF prog-id=89 op=UNLOAD Jan 24 11:41:44.286000 audit[2706]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2664 pid=2706 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.286000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6231323536363466633037653634343866613261653830306334643335 Jan 24 11:41:44.286000 audit: BPF prog-id=88 op=UNLOAD Jan 24 11:41:44.286000 audit: BPF prog-id=90 op=LOAD Jan 24 11:41:44.286000 audit[2706]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2664 pid=2706 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.286000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6231323536363466633037653634343866613261653830306334643335 Jan 24 11:41:44.292000 audit: BPF prog-id=91 op=LOAD Jan 24 11:41:44.298000 audit: BPF prog-id=92 op=LOAD Jan 24 11:41:44.298000 audit[2722]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00018c238 a2=98 a3=0 items=0 ppid=2696 pid=2722 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.298000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6139343434333237383336656564633933376436653563363034343231 Jan 24 11:41:44.298000 audit: BPF prog-id=92 op=UNLOAD Jan 24 11:41:44.298000 audit[2722]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2696 pid=2722 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.298000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6139343434333237383336656564633933376436653563363034343231 Jan 24 11:41:44.292000 audit[2706]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0000ea6e8 a2=98 a3=0 items=0 ppid=2664 pid=2706 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.292000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6231323536363466633037653634343866613261653830306334643335 Jan 24 11:41:44.332000 audit: BPF prog-id=93 op=LOAD Jan 24 11:41:44.332000 audit[2722]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00018c488 a2=98 a3=0 items=0 ppid=2696 pid=2722 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.332000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6139343434333237383336656564633933376436653563363034343231 Jan 24 11:41:44.334000 audit: BPF prog-id=94 op=LOAD Jan 24 11:41:44.334000 audit[2722]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00018c218 a2=98 a3=0 items=0 ppid=2696 pid=2722 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.334000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6139343434333237383336656564633933376436653563363034343231 Jan 24 11:41:44.334000 audit: BPF prog-id=94 op=UNLOAD Jan 24 11:41:44.334000 audit[2722]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2696 pid=2722 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.334000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6139343434333237383336656564633933376436653563363034343231 Jan 24 11:41:44.334000 audit: BPF prog-id=93 op=UNLOAD Jan 24 11:41:44.334000 audit[2722]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2696 pid=2722 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.334000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6139343434333237383336656564633933376436653563363034343231 Jan 24 11:41:44.335000 audit: BPF prog-id=95 op=LOAD Jan 24 11:41:44.335000 audit[2722]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00018c6e8 a2=98 a3=0 items=0 ppid=2696 pid=2722 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:44.335000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6139343434333237383336656564633933376436653563363034343231 Jan 24 11:41:45.455033 kubelet[2609]: E0124 11:41:45.454564 2609 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.26:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.26:6443: connect: connection refused" interval="7s" Jan 24 11:41:45.767133 containerd[1625]: time="2026-01-24T11:41:45.766680606Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:07ca0cbf79ad6ba9473d8e9f7715e571,Namespace:kube-system,Attempt:0,} returns sandbox id \"5e9dcb862ae1b53dbad14b8ca80587e8a60dfc5660e5c91d8bda9d385a7a887e\"" Jan 24 11:41:45.932019 kubelet[2609]: E0124 11:41:45.930647 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:45.944873 containerd[1625]: time="2026-01-24T11:41:45.944827943Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:5bbfee13ce9e07281eca876a0b8067f2,Namespace:kube-system,Attempt:0,} returns sandbox id \"a9444327836eedc937d6e5c604421e4e8b6cf27385034aebf9af40ac694b8db9\"" Jan 24 11:41:45.959005 kubelet[2609]: E0124 11:41:45.956889 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:45.986165 containerd[1625]: time="2026-01-24T11:41:45.985892285Z" level=info msg="CreateContainer within sandbox \"5e9dcb862ae1b53dbad14b8ca80587e8a60dfc5660e5c91d8bda9d385a7a887e\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 24 11:41:45.996807 containerd[1625]: time="2026-01-24T11:41:45.996408037Z" level=info msg="CreateContainer within sandbox \"a9444327836eedc937d6e5c604421e4e8b6cf27385034aebf9af40ac694b8db9\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 24 11:41:46.011839 containerd[1625]: time="2026-01-24T11:41:46.011695796Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:a1c0b96845ba45e5728846e5b1af0a4c,Namespace:kube-system,Attempt:0,} returns sandbox id \"b125664fc07e6448fa2ae800c4d35891c37744064bcd79d98616ad893b99d3fe\"" Jan 24 11:41:46.018792 kubelet[2609]: E0124 11:41:46.017680 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:46.042888 containerd[1625]: time="2026-01-24T11:41:46.042644381Z" level=info msg="CreateContainer within sandbox \"b125664fc07e6448fa2ae800c4d35891c37744064bcd79d98616ad893b99d3fe\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 24 11:41:46.108862 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1847793329.mount: Deactivated successfully. Jan 24 11:41:46.642153 containerd[1625]: time="2026-01-24T11:41:46.641861646Z" level=info msg="Container 1b73128ffdfcf6c526a25b3725dcc59aa9cfb85b2acbb2afe54a2332e1f6b2df: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:41:46.644670 kubelet[2609]: E0124 11:41:46.644632 2609 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.26:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 24 11:41:46.683612 containerd[1625]: time="2026-01-24T11:41:46.682477045Z" level=info msg="Container f390d4a610597155b424f663d61bff1e14a007f7da9c45b151b3cff4da3dd493: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:41:46.758769 containerd[1625]: time="2026-01-24T11:41:46.758708967Z" level=info msg="Container a5e93bdfaf18af43ae7ec55263a07cba1cab87e6ec6f22fb1ab8585874d9e37f: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:41:46.782871 containerd[1625]: time="2026-01-24T11:41:46.782817817Z" level=info msg="CreateContainer within sandbox \"5e9dcb862ae1b53dbad14b8ca80587e8a60dfc5660e5c91d8bda9d385a7a887e\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"1b73128ffdfcf6c526a25b3725dcc59aa9cfb85b2acbb2afe54a2332e1f6b2df\"" Jan 24 11:41:46.790575 containerd[1625]: time="2026-01-24T11:41:46.789756440Z" level=info msg="StartContainer for \"1b73128ffdfcf6c526a25b3725dcc59aa9cfb85b2acbb2afe54a2332e1f6b2df\"" Jan 24 11:41:46.856896 containerd[1625]: time="2026-01-24T11:41:46.856830958Z" level=info msg="CreateContainer within sandbox \"b125664fc07e6448fa2ae800c4d35891c37744064bcd79d98616ad893b99d3fe\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"a5e93bdfaf18af43ae7ec55263a07cba1cab87e6ec6f22fb1ab8585874d9e37f\"" Jan 24 11:41:46.883522 containerd[1625]: time="2026-01-24T11:41:46.882721403Z" level=info msg="CreateContainer within sandbox \"a9444327836eedc937d6e5c604421e4e8b6cf27385034aebf9af40ac694b8db9\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"f390d4a610597155b424f663d61bff1e14a007f7da9c45b151b3cff4da3dd493\"" Jan 24 11:41:46.884755 containerd[1625]: time="2026-01-24T11:41:46.884639214Z" level=info msg="StartContainer for \"f390d4a610597155b424f663d61bff1e14a007f7da9c45b151b3cff4da3dd493\"" Jan 24 11:41:46.885555 containerd[1625]: time="2026-01-24T11:41:46.885420076Z" level=info msg="StartContainer for \"a5e93bdfaf18af43ae7ec55263a07cba1cab87e6ec6f22fb1ab8585874d9e37f\"" Jan 24 11:41:46.893594 containerd[1625]: time="2026-01-24T11:41:46.893167526Z" level=info msg="connecting to shim a5e93bdfaf18af43ae7ec55263a07cba1cab87e6ec6f22fb1ab8585874d9e37f" address="unix:///run/containerd/s/afd1f8e3e32a97be4a50ecf770aeae76fb4ba5dee76bba171604355827407b49" protocol=ttrpc version=3 Jan 24 11:41:46.905580 containerd[1625]: time="2026-01-24T11:41:46.904803104Z" level=info msg="connecting to shim f390d4a610597155b424f663d61bff1e14a007f7da9c45b151b3cff4da3dd493" address="unix:///run/containerd/s/6953372d851b3c59b64a6d91f6389ab8cd31e8d7231f7e148621e804f1dac107" protocol=ttrpc version=3 Jan 24 11:41:46.911546 containerd[1625]: time="2026-01-24T11:41:46.910711782Z" level=info msg="connecting to shim 1b73128ffdfcf6c526a25b3725dcc59aa9cfb85b2acbb2afe54a2332e1f6b2df" address="unix:///run/containerd/s/486e5a14bdfba552debb04825c3a379ed542663b09e58d94f194a16faeff4d1d" protocol=ttrpc version=3 Jan 24 11:41:47.092360 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2164732207.mount: Deactivated successfully. Jan 24 11:41:47.154595 systemd[1]: Started cri-containerd-f390d4a610597155b424f663d61bff1e14a007f7da9c45b151b3cff4da3dd493.scope - libcontainer container f390d4a610597155b424f663d61bff1e14a007f7da9c45b151b3cff4da3dd493. Jan 24 11:41:47.445570 systemd[1]: Started cri-containerd-1b73128ffdfcf6c526a25b3725dcc59aa9cfb85b2acbb2afe54a2332e1f6b2df.scope - libcontainer container 1b73128ffdfcf6c526a25b3725dcc59aa9cfb85b2acbb2afe54a2332e1f6b2df. Jan 24 11:41:47.482900 systemd[1]: Started cri-containerd-a5e93bdfaf18af43ae7ec55263a07cba1cab87e6ec6f22fb1ab8585874d9e37f.scope - libcontainer container a5e93bdfaf18af43ae7ec55263a07cba1cab87e6ec6f22fb1ab8585874d9e37f. Jan 24 11:41:47.530000 audit: BPF prog-id=96 op=LOAD Jan 24 11:41:47.551846 kernel: kauditd_printk_skb: 56 callbacks suppressed Jan 24 11:41:47.552390 kernel: audit: type=1334 audit(1769254907.530:380): prog-id=96 op=LOAD Jan 24 11:41:47.552435 kubelet[2609]: I0124 11:41:47.547530 2609 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:41:47.595886 kernel: audit: type=1334 audit(1769254907.539:381): prog-id=97 op=LOAD Jan 24 11:41:47.689935 kernel: audit: type=1300 audit(1769254907.539:381): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.691804 kernel: audit: type=1327 audit(1769254907.539:381): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:47.692145 kernel: audit: type=1334 audit(1769254907.539:382): prog-id=97 op=UNLOAD Jan 24 11:41:47.832119 kernel: audit: type=1300 audit(1769254907.539:382): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.832466 kernel: audit: type=1327 audit(1769254907.539:382): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:47.832753 kernel: audit: type=1334 audit(1769254907.539:383): prog-id=98 op=LOAD Jan 24 11:41:47.880730 kernel: audit: type=1300 audit(1769254907.539:383): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.539000 audit: BPF prog-id=97 op=LOAD Jan 24 11:41:47.539000 audit[2793]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.539000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:47.539000 audit: BPF prog-id=97 op=UNLOAD Jan 24 11:41:47.539000 audit[2793]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.539000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:47.539000 audit: BPF prog-id=98 op=LOAD Jan 24 11:41:47.539000 audit[2793]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.539000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:47.959004 kubelet[2609]: E0124 11:41:47.613887 2609 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.26:6443/api/v1/nodes\": dial tcp 10.0.0.26:6443: connect: connection refused" node="localhost" Jan 24 11:41:48.062492 kernel: audit: type=1327 audit(1769254907.539:383): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:47.540000 audit: BPF prog-id=99 op=LOAD Jan 24 11:41:47.540000 audit[2793]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.540000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:47.540000 audit: BPF prog-id=99 op=UNLOAD Jan 24 11:41:47.540000 audit[2793]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.540000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:47.540000 audit: BPF prog-id=98 op=UNLOAD Jan 24 11:41:47.540000 audit[2793]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.540000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:47.540000 audit: BPF prog-id=100 op=LOAD Jan 24 11:41:47.540000 audit[2793]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2696 pid=2793 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:47.540000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6633393064346136313035393731353562343234663636336436316266 Jan 24 11:41:48.184609 kubelet[2609]: E0124 11:41:48.184560 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.26:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 24 11:41:48.196000 audit: BPF prog-id=101 op=LOAD Jan 24 11:41:48.207000 audit: BPF prog-id=102 op=LOAD Jan 24 11:41:48.207000 audit[2794]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2664 pid=2794 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.207000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6135653933626466616631386166343361653765633535323633613037 Jan 24 11:41:48.208000 audit: BPF prog-id=102 op=UNLOAD Jan 24 11:41:48.208000 audit[2794]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2664 pid=2794 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.208000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6135653933626466616631386166343361653765633535323633613037 Jan 24 11:41:48.208000 audit: BPF prog-id=103 op=LOAD Jan 24 11:41:48.208000 audit[2794]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2664 pid=2794 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.208000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6135653933626466616631386166343361653765633535323633613037 Jan 24 11:41:48.209000 audit: BPF prog-id=104 op=LOAD Jan 24 11:41:48.209000 audit[2794]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2664 pid=2794 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.209000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6135653933626466616631386166343361653765633535323633613037 Jan 24 11:41:48.209000 audit: BPF prog-id=104 op=UNLOAD Jan 24 11:41:48.209000 audit[2794]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2664 pid=2794 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.209000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6135653933626466616631386166343361653765633535323633613037 Jan 24 11:41:48.209000 audit: BPF prog-id=103 op=UNLOAD Jan 24 11:41:48.209000 audit[2794]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2664 pid=2794 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.209000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6135653933626466616631386166343361653765633535323633613037 Jan 24 11:41:48.242000 audit: BPF prog-id=105 op=LOAD Jan 24 11:41:48.242000 audit[2794]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2664 pid=2794 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.242000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6135653933626466616631386166343361653765633535323633613037 Jan 24 11:41:48.280000 audit: BPF prog-id=106 op=LOAD Jan 24 11:41:48.283000 audit: BPF prog-id=107 op=LOAD Jan 24 11:41:48.283000 audit[2795]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2676 pid=2795 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373331323866666466636636633532366132356233373235646363 Jan 24 11:41:48.283000 audit: BPF prog-id=107 op=UNLOAD Jan 24 11:41:48.283000 audit[2795]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2676 pid=2795 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373331323866666466636636633532366132356233373235646363 Jan 24 11:41:48.283000 audit: BPF prog-id=108 op=LOAD Jan 24 11:41:48.283000 audit[2795]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2676 pid=2795 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373331323866666466636636633532366132356233373235646363 Jan 24 11:41:48.283000 audit: BPF prog-id=109 op=LOAD Jan 24 11:41:48.283000 audit[2795]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2676 pid=2795 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373331323866666466636636633532366132356233373235646363 Jan 24 11:41:48.283000 audit: BPF prog-id=109 op=UNLOAD Jan 24 11:41:48.283000 audit[2795]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2676 pid=2795 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373331323866666466636636633532366132356233373235646363 Jan 24 11:41:48.283000 audit: BPF prog-id=108 op=UNLOAD Jan 24 11:41:48.283000 audit[2795]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2676 pid=2795 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373331323866666466636636633532366132356233373235646363 Jan 24 11:41:48.283000 audit: BPF prog-id=110 op=LOAD Jan 24 11:41:48.283000 audit[2795]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2676 pid=2795 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:41:48.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373331323866666466636636633532366132356233373235646363 Jan 24 11:41:48.666450 containerd[1625]: time="2026-01-24T11:41:48.666308111Z" level=info msg="StartContainer for \"a5e93bdfaf18af43ae7ec55263a07cba1cab87e6ec6f22fb1ab8585874d9e37f\" returns successfully" Jan 24 11:41:48.761867 containerd[1625]: time="2026-01-24T11:41:48.758970746Z" level=info msg="StartContainer for \"1b73128ffdfcf6c526a25b3725dcc59aa9cfb85b2acbb2afe54a2332e1f6b2df\" returns successfully" Jan 24 11:41:48.762501 containerd[1625]: time="2026-01-24T11:41:48.760349011Z" level=info msg="StartContainer for \"f390d4a610597155b424f663d61bff1e14a007f7da9c45b151b3cff4da3dd493\" returns successfully" Jan 24 11:41:49.585729 kubelet[2609]: E0124 11:41:49.585512 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.26:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 24 11:41:49.661337 kubelet[2609]: E0124 11:41:49.660950 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:49.661811 kubelet[2609]: E0124 11:41:49.661784 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:49.691905 kubelet[2609]: E0124 11:41:49.689917 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:49.703526 kubelet[2609]: E0124 11:41:49.697914 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:49.768459 kubelet[2609]: E0124 11:41:49.750653 2609 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.26:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.26:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.188da7f366a4ec6b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-24 11:41:32.467448939 +0000 UTC m=+2.736922269,LastTimestamp:2026-01-24 11:41:32.467448939 +0000 UTC m=+2.736922269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 24 11:41:49.785481 kubelet[2609]: E0124 11:41:49.784647 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:49.785481 kubelet[2609]: E0124 11:41:49.784935 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:50.635454 kubelet[2609]: E0124 11:41:50.634600 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.26:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.26:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 24 11:41:51.053814 kubelet[2609]: E0124 11:41:51.049568 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:51.063117 kubelet[2609]: E0124 11:41:51.057808 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:51.063117 kubelet[2609]: E0124 11:41:51.056393 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:51.063117 kubelet[2609]: E0124 11:41:51.058335 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:51.063117 kubelet[2609]: E0124 11:41:51.051892 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:51.063117 kubelet[2609]: E0124 11:41:51.059472 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:52.064464 kubelet[2609]: E0124 11:41:52.063134 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:52.064464 kubelet[2609]: E0124 11:41:52.062641 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:52.064464 kubelet[2609]: E0124 11:41:52.063741 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:52.067406 kubelet[2609]: E0124 11:41:52.065388 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:54.124446 kubelet[2609]: E0124 11:41:54.123364 2609 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 24 11:41:54.642069 kubelet[2609]: I0124 11:41:54.641661 2609 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:41:56.190611 kubelet[2609]: E0124 11:41:56.188857 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:56.194782 kubelet[2609]: E0124 11:41:56.191823 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:56.752058 kubelet[2609]: E0124 11:41:56.751596 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:56.782700 kubelet[2609]: E0124 11:41:56.780886 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:41:59.180065 kubelet[2609]: E0124 11:41:59.179488 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:41:59.183703 kubelet[2609]: E0124 11:41:59.180866 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:02.140708 kubelet[2609]: E0124 11:42:02.139672 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.26:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 24 11:42:02.463512 kubelet[2609]: E0124 11:42:02.461596 2609 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.26:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="7s" Jan 24 11:42:04.125626 kubelet[2609]: E0124 11:42:04.125364 2609 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 24 11:42:04.646721 kubelet[2609]: E0124 11:42:04.646101 2609 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.26:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="localhost" Jan 24 11:42:06.150705 kubelet[2609]: E0124 11:42:06.149745 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:42:06.155082 kubelet[2609]: E0124 11:42:06.153454 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:10.031085 kubelet[2609]: E0124 11:42:10.028866 2609 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.26:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{localhost.188da7f366a4ec6b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-24 11:41:32.467448939 +0000 UTC m=+2.736922269,LastTimestamp:2026-01-24 11:41:32.467448939 +0000 UTC m=+2.736922269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 24 11:42:11.660089 kubelet[2609]: I0124 11:42:11.659362 2609 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:42:13.991495 kubelet[2609]: E0124 11:42:13.988461 2609 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.26:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 24 11:42:14.113783 kubelet[2609]: E0124 11:42:13.989099 2609 certificate_manager.go:461] "Reached backoff limit, still unable to rotate certs" err="timed out waiting for the condition" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 24 11:42:14.127848 kubelet[2609]: E0124 11:42:14.127762 2609 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 24 11:42:14.299137 kubelet[2609]: E0124 11:42:14.296596 2609 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:42:14.324776 kubelet[2609]: E0124 11:42:14.323811 2609 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:15.040106 kubelet[2609]: E0124 11:42:15.038871 2609 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.26:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 24 11:42:16.574303 kubelet[2609]: E0124 11:42:16.573688 2609 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Jan 24 11:42:16.772329 kubelet[2609]: I0124 11:42:16.770945 2609 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jan 24 11:42:16.786832 kubelet[2609]: E0124 11:42:16.784402 2609 kubelet_node_status.go:486] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Jan 24 11:42:16.982654 kubelet[2609]: E0124 11:42:16.982048 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:17.140558 kubelet[2609]: E0124 11:42:17.138899 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:17.243286 kubelet[2609]: E0124 11:42:17.240376 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:17.341428 kubelet[2609]: E0124 11:42:17.341371 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:17.462061 kubelet[2609]: E0124 11:42:17.459570 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:17.598858 kubelet[2609]: E0124 11:42:17.574606 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:17.692878 kubelet[2609]: E0124 11:42:17.690021 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:17.791911 kubelet[2609]: E0124 11:42:17.790994 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:17.895317 kubelet[2609]: E0124 11:42:17.892313 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:17.995729 kubelet[2609]: E0124 11:42:17.995490 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:18.128347 kubelet[2609]: E0124 11:42:18.127000 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:18.367683 kubelet[2609]: E0124 11:42:18.243779 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:18.418792 kubelet[2609]: E0124 11:42:18.395926 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:18.528620 kubelet[2609]: E0124 11:42:18.527122 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:18.636720 kubelet[2609]: E0124 11:42:18.632819 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:18.749387 kubelet[2609]: E0124 11:42:18.749320 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:18.850855 kubelet[2609]: E0124 11:42:18.850635 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.147081 kubelet[2609]: E0124 11:42:19.028834 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.147081 kubelet[2609]: E0124 11:42:19.144536 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.246884 kubelet[2609]: E0124 11:42:19.246075 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.350845 kubelet[2609]: E0124 11:42:19.350587 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.459328 kubelet[2609]: E0124 11:42:19.457657 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.561381 kubelet[2609]: E0124 11:42:19.560307 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.671026 kubelet[2609]: E0124 11:42:19.663550 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.769922 kubelet[2609]: E0124 11:42:19.769479 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.870606 kubelet[2609]: E0124 11:42:19.870540 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:19.972374 kubelet[2609]: E0124 11:42:19.971863 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.074616 kubelet[2609]: E0124 11:42:20.074159 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.174478 kubelet[2609]: E0124 11:42:20.174401 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.278284 kubelet[2609]: E0124 11:42:20.277629 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.380082 kubelet[2609]: E0124 11:42:20.378053 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.479461 kubelet[2609]: E0124 11:42:20.479308 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.580526 kubelet[2609]: E0124 11:42:20.580160 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.680789 kubelet[2609]: E0124 11:42:20.680639 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.780870 kubelet[2609]: E0124 11:42:20.780779 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.884485 kubelet[2609]: E0124 11:42:20.884367 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:20.989057 kubelet[2609]: E0124 11:42:20.985129 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:21.090302 kubelet[2609]: E0124 11:42:21.088535 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:21.190383 kubelet[2609]: E0124 11:42:21.189288 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:21.290799 kubelet[2609]: E0124 11:42:21.290272 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:21.393148 kubelet[2609]: E0124 11:42:21.391844 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:21.494788 kubelet[2609]: E0124 11:42:21.494505 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:21.595550 kubelet[2609]: E0124 11:42:21.595170 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:21.697554 kubelet[2609]: E0124 11:42:21.697030 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:21.820618 kubelet[2609]: E0124 11:42:21.798177 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:21.916741 kubelet[2609]: E0124 11:42:21.912632 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.018097 kubelet[2609]: E0124 11:42:22.017104 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.127450 kubelet[2609]: E0124 11:42:22.124057 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.231580 kubelet[2609]: E0124 11:42:22.231414 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.333281 kubelet[2609]: E0124 11:42:22.331927 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.439290 kubelet[2609]: E0124 11:42:22.438694 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.540366 kubelet[2609]: E0124 11:42:22.539903 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.640650 kubelet[2609]: E0124 11:42:22.640585 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.743837 kubelet[2609]: E0124 11:42:22.743400 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.777846 systemd[1]: Reload requested from client PID 2906 ('systemctl') (unit session-10.scope)... Jan 24 11:42:22.778602 systemd[1]: Reloading... Jan 24 11:42:22.848305 kubelet[2609]: E0124 11:42:22.847979 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:22.948474 kubelet[2609]: E0124 11:42:22.948390 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:23.051820 kubelet[2609]: E0124 11:42:23.048970 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:23.059011 zram_generator::config[2960]: No configuration found. Jan 24 11:42:23.150418 kubelet[2609]: E0124 11:42:23.149417 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:23.251048 kubelet[2609]: E0124 11:42:23.250786 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:23.352987 kubelet[2609]: E0124 11:42:23.352621 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:23.455844 kubelet[2609]: E0124 11:42:23.455704 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:23.558592 kubelet[2609]: E0124 11:42:23.557722 2609 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:42:23.568663 systemd[1]: Reloading finished in 789 ms. Jan 24 11:42:23.641890 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:42:23.658132 systemd[1]: kubelet.service: Deactivated successfully. Jan 24 11:42:23.658680 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:42:23.658000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:42:23.658792 systemd[1]: kubelet.service: Consumed 12.140s CPU time, 128.8M memory peak. Jan 24 11:42:23.661887 kernel: kauditd_printk_skb: 56 callbacks suppressed Jan 24 11:42:23.662152 kernel: audit: type=1131 audit(1769254943.658:404): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:42:23.664385 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:42:23.668000 audit: BPF prog-id=111 op=LOAD Jan 24 11:42:23.681588 kernel: audit: type=1334 audit(1769254943.668:405): prog-id=111 op=LOAD Jan 24 11:42:23.681684 kernel: audit: type=1334 audit(1769254943.668:406): prog-id=67 op=UNLOAD Jan 24 11:42:23.668000 audit: BPF prog-id=67 op=UNLOAD Jan 24 11:42:23.672000 audit: BPF prog-id=112 op=LOAD Jan 24 11:42:23.688601 kernel: audit: type=1334 audit(1769254943.672:407): prog-id=112 op=LOAD Jan 24 11:42:23.672000 audit: BPF prog-id=68 op=UNLOAD Jan 24 11:42:23.689330 kernel: audit: type=1334 audit(1769254943.672:408): prog-id=68 op=UNLOAD Jan 24 11:42:23.715828 kernel: audit: type=1334 audit(1769254943.672:409): prog-id=113 op=LOAD Jan 24 11:42:23.730767 kernel: audit: type=1334 audit(1769254943.673:410): prog-id=114 op=LOAD Jan 24 11:42:23.672000 audit: BPF prog-id=113 op=LOAD Jan 24 11:42:23.673000 audit: BPF prog-id=114 op=LOAD Jan 24 11:42:23.673000 audit: BPF prog-id=69 op=UNLOAD Jan 24 11:42:23.740261 kernel: audit: type=1334 audit(1769254943.673:411): prog-id=69 op=UNLOAD Jan 24 11:42:23.673000 audit: BPF prog-id=70 op=UNLOAD Jan 24 11:42:23.748133 kernel: audit: type=1334 audit(1769254943.673:412): prog-id=70 op=UNLOAD Jan 24 11:42:23.748334 kernel: audit: type=1334 audit(1769254943.674:413): prog-id=115 op=LOAD Jan 24 11:42:23.674000 audit: BPF prog-id=115 op=LOAD Jan 24 11:42:23.674000 audit: BPF prog-id=77 op=UNLOAD Jan 24 11:42:23.674000 audit: BPF prog-id=116 op=LOAD Jan 24 11:42:23.675000 audit: BPF prog-id=117 op=LOAD Jan 24 11:42:23.675000 audit: BPF prog-id=62 op=UNLOAD Jan 24 11:42:23.675000 audit: BPF prog-id=63 op=UNLOAD Jan 24 11:42:23.679000 audit: BPF prog-id=118 op=LOAD Jan 24 11:42:23.679000 audit: BPF prog-id=64 op=UNLOAD Jan 24 11:42:23.680000 audit: BPF prog-id=119 op=LOAD Jan 24 11:42:23.680000 audit: BPF prog-id=120 op=LOAD Jan 24 11:42:23.680000 audit: BPF prog-id=65 op=UNLOAD Jan 24 11:42:23.680000 audit: BPF prog-id=66 op=UNLOAD Jan 24 11:42:23.698000 audit: BPF prog-id=121 op=LOAD Jan 24 11:42:23.698000 audit: BPF prog-id=71 op=UNLOAD Jan 24 11:42:23.728000 audit: BPF prog-id=122 op=LOAD Jan 24 11:42:23.728000 audit: BPF prog-id=123 op=LOAD Jan 24 11:42:23.728000 audit: BPF prog-id=72 op=UNLOAD Jan 24 11:42:23.728000 audit: BPF prog-id=73 op=UNLOAD Jan 24 11:42:23.732000 audit: BPF prog-id=124 op=LOAD Jan 24 11:42:23.732000 audit: BPF prog-id=61 op=UNLOAD Jan 24 11:42:23.734000 audit: BPF prog-id=125 op=LOAD Jan 24 11:42:23.734000 audit: BPF prog-id=74 op=UNLOAD Jan 24 11:42:23.734000 audit: BPF prog-id=126 op=LOAD Jan 24 11:42:23.734000 audit: BPF prog-id=127 op=LOAD Jan 24 11:42:23.734000 audit: BPF prog-id=75 op=UNLOAD Jan 24 11:42:23.734000 audit: BPF prog-id=76 op=UNLOAD Jan 24 11:42:23.737000 audit: BPF prog-id=128 op=LOAD Jan 24 11:42:23.737000 audit: BPF prog-id=78 op=UNLOAD Jan 24 11:42:23.737000 audit: BPF prog-id=129 op=LOAD Jan 24 11:42:23.737000 audit: BPF prog-id=130 op=LOAD Jan 24 11:42:23.737000 audit: BPF prog-id=79 op=UNLOAD Jan 24 11:42:23.737000 audit: BPF prog-id=80 op=UNLOAD Jan 24 11:42:24.158000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:42:24.159623 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:42:24.180808 (kubelet)[2999]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 24 11:42:24.353537 kubelet[2999]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 24 11:42:24.353537 kubelet[2999]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 24 11:42:24.353537 kubelet[2999]: I0124 11:42:24.352859 2999 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 24 11:42:24.387639 kubelet[2999]: I0124 11:42:24.387588 2999 server.go:529] "Kubelet version" kubeletVersion="v1.34.1" Jan 24 11:42:24.388083 kubelet[2999]: I0124 11:42:24.388061 2999 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 24 11:42:24.389500 kubelet[2999]: I0124 11:42:24.388333 2999 watchdog_linux.go:95] "Systemd watchdog is not enabled" Jan 24 11:42:24.389500 kubelet[2999]: I0124 11:42:24.388389 2999 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 24 11:42:24.389500 kubelet[2999]: I0124 11:42:24.388837 2999 server.go:956] "Client rotation is on, will bootstrap in background" Jan 24 11:42:24.390728 kubelet[2999]: I0124 11:42:24.390640 2999 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Jan 24 11:42:24.396284 kubelet[2999]: I0124 11:42:24.394862 2999 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 24 11:42:24.419644 kubelet[2999]: I0124 11:42:24.419361 2999 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 24 11:42:24.444470 kubelet[2999]: I0124 11:42:24.443469 2999 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Jan 24 11:42:24.444470 kubelet[2999]: I0124 11:42:24.443900 2999 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 24 11:42:24.444470 kubelet[2999]: I0124 11:42:24.444053 2999 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 24 11:42:24.444470 kubelet[2999]: I0124 11:42:24.444460 2999 topology_manager.go:138] "Creating topology manager with none policy" Jan 24 11:42:24.444865 kubelet[2999]: I0124 11:42:24.444478 2999 container_manager_linux.go:306] "Creating device plugin manager" Jan 24 11:42:24.444865 kubelet[2999]: I0124 11:42:24.444516 2999 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Jan 24 11:42:24.446776 kubelet[2999]: I0124 11:42:24.446539 2999 state_mem.go:36] "Initialized new in-memory state store" Jan 24 11:42:24.452290 kubelet[2999]: I0124 11:42:24.448343 2999 kubelet.go:475] "Attempting to sync node with API server" Jan 24 11:42:24.452290 kubelet[2999]: I0124 11:42:24.448366 2999 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 24 11:42:24.452290 kubelet[2999]: I0124 11:42:24.448395 2999 kubelet.go:387] "Adding apiserver pod source" Jan 24 11:42:24.452290 kubelet[2999]: I0124 11:42:24.448427 2999 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 24 11:42:24.474865 kubelet[2999]: I0124 11:42:24.474735 2999 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 24 11:42:24.491632 kubelet[2999]: I0124 11:42:24.490308 2999 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 24 11:42:24.492243 kubelet[2999]: I0124 11:42:24.492147 2999 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Jan 24 11:42:24.527370 kubelet[2999]: I0124 11:42:24.522062 2999 server.go:1262] "Started kubelet" Jan 24 11:42:24.527370 kubelet[2999]: I0124 11:42:24.525384 2999 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 24 11:42:24.527370 kubelet[2999]: I0124 11:42:24.525988 2999 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 24 11:42:24.527370 kubelet[2999]: I0124 11:42:24.526165 2999 server_v1.go:49] "podresources" method="list" useActivePods=true Jan 24 11:42:24.531315 kubelet[2999]: I0124 11:42:24.529009 2999 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 24 11:42:24.531315 kubelet[2999]: I0124 11:42:24.529828 2999 volume_manager.go:313] "Starting Kubelet Volume Manager" Jan 24 11:42:24.537986 kubelet[2999]: I0124 11:42:24.523655 2999 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 24 11:42:24.537986 kubelet[2999]: I0124 11:42:24.537260 2999 server.go:310] "Adding debug handlers to kubelet server" Jan 24 11:42:24.542136 kubelet[2999]: I0124 11:42:24.540423 2999 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 24 11:42:24.542136 kubelet[2999]: I0124 11:42:24.540666 2999 reconciler.go:29] "Reconciler: start to sync state" Jan 24 11:42:24.542136 kubelet[2999]: I0124 11:42:24.540877 2999 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 24 11:42:24.557113 kubelet[2999]: I0124 11:42:24.555810 2999 factory.go:223] Registration of the systemd container factory successfully Jan 24 11:42:24.557113 kubelet[2999]: I0124 11:42:24.556079 2999 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 24 11:42:24.567665 kubelet[2999]: I0124 11:42:24.567586 2999 factory.go:223] Registration of the containerd container factory successfully Jan 24 11:42:24.568781 kubelet[2999]: E0124 11:42:24.568700 2999 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 24 11:42:24.657167 kubelet[2999]: I0124 11:42:24.653138 2999 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Jan 24 11:42:24.657167 kubelet[2999]: I0124 11:42:24.656733 2999 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Jan 24 11:42:24.663255 kubelet[2999]: I0124 11:42:24.659548 2999 status_manager.go:244] "Starting to sync pod status with apiserver" Jan 24 11:42:24.663255 kubelet[2999]: I0124 11:42:24.659631 2999 kubelet.go:2427] "Starting kubelet main sync loop" Jan 24 11:42:24.663255 kubelet[2999]: E0124 11:42:24.659703 2999 kubelet.go:2451] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746160 2999 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746276 2999 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746304 2999 state_mem.go:36] "Initialized new in-memory state store" Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746479 2999 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746502 2999 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746527 2999 policy_none.go:49] "None policy: Start" Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746544 2999 memory_manager.go:187] "Starting memorymanager" policy="None" Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746561 2999 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746689 2999 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Jan 24 11:42:24.747811 kubelet[2999]: I0124 11:42:24.746702 2999 policy_none.go:47] "Start" Jan 24 11:42:24.760395 kubelet[2999]: E0124 11:42:24.760343 2999 kubelet.go:2451] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 24 11:42:24.774519 kubelet[2999]: E0124 11:42:24.774113 2999 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 24 11:42:24.774519 kubelet[2999]: I0124 11:42:24.774453 2999 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 24 11:42:24.777574 kubelet[2999]: I0124 11:42:24.774469 2999 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 24 11:42:24.778778 kubelet[2999]: I0124 11:42:24.778595 2999 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 24 11:42:24.781436 kubelet[2999]: E0124 11:42:24.780547 2999 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 24 11:42:24.921912 kubelet[2999]: I0124 11:42:24.921620 2999 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:42:24.973742 kubelet[2999]: I0124 11:42:24.972428 2999 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jan 24 11:42:24.973742 kubelet[2999]: I0124 11:42:24.972439 2999 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jan 24 11:42:24.991741 kubelet[2999]: I0124 11:42:24.986610 2999 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 24 11:42:24.995606 kubelet[2999]: I0124 11:42:24.994598 2999 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Jan 24 11:42:24.995606 kubelet[2999]: I0124 11:42:24.995290 2999 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jan 24 11:42:25.051641 kubelet[2999]: I0124 11:42:25.049632 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:42:25.051641 kubelet[2999]: I0124 11:42:25.049739 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/07ca0cbf79ad6ba9473d8e9f7715e571-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"07ca0cbf79ad6ba9473d8e9f7715e571\") " pod="kube-system/kube-scheduler-localhost" Jan 24 11:42:25.051641 kubelet[2999]: I0124 11:42:25.049772 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:42:25.051641 kubelet[2999]: I0124 11:42:25.049798 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:42:25.051641 kubelet[2999]: I0124 11:42:25.049825 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a1c0b96845ba45e5728846e5b1af0a4c-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"a1c0b96845ba45e5728846e5b1af0a4c\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:42:25.051996 kubelet[2999]: I0124 11:42:25.049844 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a1c0b96845ba45e5728846e5b1af0a4c-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"a1c0b96845ba45e5728846e5b1af0a4c\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:42:25.051996 kubelet[2999]: I0124 11:42:25.049875 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a1c0b96845ba45e5728846e5b1af0a4c-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"a1c0b96845ba45e5728846e5b1af0a4c\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:42:25.051996 kubelet[2999]: I0124 11:42:25.049900 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:42:25.051996 kubelet[2999]: I0124 11:42:25.049979 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:42:25.437679 kubelet[2999]: E0124 11:42:25.437545 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:25.447505 kubelet[2999]: E0124 11:42:25.445346 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:25.447505 kubelet[2999]: E0124 11:42:25.445584 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:25.456062 kubelet[2999]: I0124 11:42:25.455486 2999 apiserver.go:52] "Watching apiserver" Jan 24 11:42:25.546779 kubelet[2999]: I0124 11:42:25.546688 2999 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 24 11:42:25.747740 kubelet[2999]: E0124 11:42:25.741287 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:25.747740 kubelet[2999]: E0124 11:42:25.743475 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:25.747740 kubelet[2999]: I0124 11:42:25.743553 2999 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jan 24 11:42:25.779844 kubelet[2999]: E0124 11:42:25.779742 2999 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Jan 24 11:42:25.781842 kubelet[2999]: E0124 11:42:25.781763 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:25.873526 kubelet[2999]: I0124 11:42:25.873151 2999 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.873057132 podStartE2EDuration="1.873057132s" podCreationTimestamp="2026-01-24 11:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:42:25.868499422 +0000 UTC m=+1.667726619" watchObservedRunningTime="2026-01-24 11:42:25.873057132 +0000 UTC m=+1.672284319" Jan 24 11:42:25.963680 kubelet[2999]: I0124 11:42:25.963556 2999 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=0.963534897 podStartE2EDuration="963.534897ms" podCreationTimestamp="2026-01-24 11:42:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:42:25.923794303 +0000 UTC m=+1.723021500" watchObservedRunningTime="2026-01-24 11:42:25.963534897 +0000 UTC m=+1.762762094" Jan 24 11:42:25.964831 kubelet[2999]: I0124 11:42:25.963693 2999 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.9636844660000001 podStartE2EDuration="1.963684466s" podCreationTimestamp="2026-01-24 11:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:42:25.961386034 +0000 UTC m=+1.760613222" watchObservedRunningTime="2026-01-24 11:42:25.963684466 +0000 UTC m=+1.762911663" Jan 24 11:42:26.745492 kubelet[2999]: E0124 11:42:26.744343 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:26.745492 kubelet[2999]: E0124 11:42:26.744458 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:26.745492 kubelet[2999]: E0124 11:42:26.745006 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:27.751445 kubelet[2999]: E0124 11:42:27.749254 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:27.794674 kubelet[2999]: E0124 11:42:27.791332 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:28.043874 kubelet[2999]: E0124 11:42:28.039627 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:28.045500 kubelet[2999]: I0124 11:42:28.044842 2999 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 24 11:42:28.057135 containerd[1625]: time="2026-01-24T11:42:28.056721872Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 24 11:42:28.068073 kubelet[2999]: I0124 11:42:28.057444 2999 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 24 11:42:28.833693 kubelet[2999]: E0124 11:42:28.833452 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:28.930523 systemd[1]: Created slice kubepods-besteffort-pod954bc09d_ef76_4327_b51f_982c86006b9b.slice - libcontainer container kubepods-besteffort-pod954bc09d_ef76_4327_b51f_982c86006b9b.slice. Jan 24 11:42:29.076668 kubelet[2999]: I0124 11:42:29.076399 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/954bc09d-ef76-4327-b51f-982c86006b9b-kube-proxy\") pod \"kube-proxy-4b29f\" (UID: \"954bc09d-ef76-4327-b51f-982c86006b9b\") " pod="kube-system/kube-proxy-4b29f" Jan 24 11:42:29.077945 kubelet[2999]: I0124 11:42:29.077551 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/954bc09d-ef76-4327-b51f-982c86006b9b-lib-modules\") pod \"kube-proxy-4b29f\" (UID: \"954bc09d-ef76-4327-b51f-982c86006b9b\") " pod="kube-system/kube-proxy-4b29f" Jan 24 11:42:29.079322 kubelet[2999]: I0124 11:42:29.078538 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2j7d\" (UniqueName: \"kubernetes.io/projected/954bc09d-ef76-4327-b51f-982c86006b9b-kube-api-access-b2j7d\") pod \"kube-proxy-4b29f\" (UID: \"954bc09d-ef76-4327-b51f-982c86006b9b\") " pod="kube-system/kube-proxy-4b29f" Jan 24 11:42:29.079777 kubelet[2999]: I0124 11:42:29.079654 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/954bc09d-ef76-4327-b51f-982c86006b9b-xtables-lock\") pod \"kube-proxy-4b29f\" (UID: \"954bc09d-ef76-4327-b51f-982c86006b9b\") " pod="kube-system/kube-proxy-4b29f" Jan 24 11:42:29.583406 kubelet[2999]: E0124 11:42:29.582656 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:29.587171 containerd[1625]: time="2026-01-24T11:42:29.585256925Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4b29f,Uid:954bc09d-ef76-4327-b51f-982c86006b9b,Namespace:kube-system,Attempt:0,}" Jan 24 11:42:29.673976 kubelet[2999]: I0124 11:42:29.673858 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/47852d4e-adf5-4a93-801a-1e88b7d8bc08-var-lib-calico\") pod \"tigera-operator-65cdcdfd6d-lgqpv\" (UID: \"47852d4e-adf5-4a93-801a-1e88b7d8bc08\") " pod="tigera-operator/tigera-operator-65cdcdfd6d-lgqpv" Jan 24 11:42:29.675597 kubelet[2999]: I0124 11:42:29.675448 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg5t7\" (UniqueName: \"kubernetes.io/projected/47852d4e-adf5-4a93-801a-1e88b7d8bc08-kube-api-access-hg5t7\") pod \"tigera-operator-65cdcdfd6d-lgqpv\" (UID: \"47852d4e-adf5-4a93-801a-1e88b7d8bc08\") " pod="tigera-operator/tigera-operator-65cdcdfd6d-lgqpv" Jan 24 11:42:29.688814 systemd[1]: Created slice kubepods-besteffort-pod47852d4e_adf5_4a93_801a_1e88b7d8bc08.slice - libcontainer container kubepods-besteffort-pod47852d4e_adf5_4a93_801a_1e88b7d8bc08.slice. Jan 24 11:42:29.843031 containerd[1625]: time="2026-01-24T11:42:29.842815365Z" level=info msg="connecting to shim 4235ad4becbd424b6f7278aa651792e19f50baab8066c7a5e4303717bdcf9e22" address="unix:///run/containerd/s/f521e1548f4c377709a04de02e875d100e951e5f99f9cfaf4fb06e5e703c5340" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:42:29.843171 kubelet[2999]: E0124 11:42:29.843019 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:29.964056 systemd[1]: Started cri-containerd-4235ad4becbd424b6f7278aa651792e19f50baab8066c7a5e4303717bdcf9e22.scope - libcontainer container 4235ad4becbd424b6f7278aa651792e19f50baab8066c7a5e4303717bdcf9e22. Jan 24 11:42:29.994000 audit: BPF prog-id=131 op=LOAD Jan 24 11:42:30.005449 kernel: kauditd_printk_skb: 32 callbacks suppressed Jan 24 11:42:30.005692 kernel: audit: type=1334 audit(1769254949.994:446): prog-id=131 op=LOAD Jan 24 11:42:30.001000 audit: BPF prog-id=132 op=LOAD Jan 24 11:42:30.055716 kernel: audit: type=1334 audit(1769254950.001:447): prog-id=132 op=LOAD Jan 24 11:42:30.056385 kernel: audit: type=1300 audit(1769254950.001:447): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.001000 audit[3081]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.071389 kernel: audit: type=1327 audit(1769254950.001:447): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.001000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.098862 kernel: audit: type=1334 audit(1769254950.001:448): prog-id=132 op=UNLOAD Jan 24 11:42:30.001000 audit: BPF prog-id=132 op=UNLOAD Jan 24 11:42:30.001000 audit[3081]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.115165 containerd[1625]: time="2026-01-24T11:42:30.114624489Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-65cdcdfd6d-lgqpv,Uid:47852d4e-adf5-4a93-801a-1e88b7d8bc08,Namespace:tigera-operator,Attempt:0,}" Jan 24 11:42:30.001000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.143292 kernel: audit: type=1300 audit(1769254950.001:448): arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.143991 kernel: audit: type=1327 audit(1769254950.001:448): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.144042 kernel: audit: type=1334 audit(1769254950.001:449): prog-id=133 op=LOAD Jan 24 11:42:30.001000 audit: BPF prog-id=133 op=LOAD Jan 24 11:42:30.001000 audit[3081]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.190306 kernel: audit: type=1300 audit(1769254950.001:449): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.192714 kernel: audit: type=1327 audit(1769254950.001:449): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.001000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.001000 audit: BPF prog-id=134 op=LOAD Jan 24 11:42:30.001000 audit[3081]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.001000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.001000 audit: BPF prog-id=134 op=UNLOAD Jan 24 11:42:30.001000 audit[3081]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.001000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.001000 audit: BPF prog-id=133 op=UNLOAD Jan 24 11:42:30.001000 audit[3081]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.001000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.001000 audit: BPF prog-id=135 op=LOAD Jan 24 11:42:30.001000 audit[3081]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3068 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:30.001000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432333561643462656362643432346236663732373861613635313739 Jan 24 11:42:30.387494 containerd[1625]: time="2026-01-24T11:42:30.386748264Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4b29f,Uid:954bc09d-ef76-4327-b51f-982c86006b9b,Namespace:kube-system,Attempt:0,} returns sandbox id \"4235ad4becbd424b6f7278aa651792e19f50baab8066c7a5e4303717bdcf9e22\"" Jan 24 11:42:30.392410 containerd[1625]: time="2026-01-24T11:42:30.390169106Z" level=info msg="connecting to shim 534b8923c66f2f4c81f8a584d3fa64023c4f561d2b556be903c5e17f5f031d49" address="unix:///run/containerd/s/2bcdf48c2c7ad57d1d79c02329d7eed9a4279291750057beb3562ceaabaed610" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:42:30.396553 kubelet[2999]: E0124 11:42:30.396393 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:30.438347 containerd[1625]: time="2026-01-24T11:42:30.437147929Z" level=info msg="CreateContainer within sandbox \"4235ad4becbd424b6f7278aa651792e19f50baab8066c7a5e4303717bdcf9e22\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 24 11:42:30.485784 containerd[1625]: time="2026-01-24T11:42:30.485515866Z" level=info msg="Container 59672d2115ce67dcb8ee3b08b0d3459c844255fb8bb14dbd511e4b7f35efb5c7: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:42:30.752168 containerd[1625]: time="2026-01-24T11:42:30.749813106Z" level=info msg="CreateContainer within sandbox \"4235ad4becbd424b6f7278aa651792e19f50baab8066c7a5e4303717bdcf9e22\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"59672d2115ce67dcb8ee3b08b0d3459c844255fb8bb14dbd511e4b7f35efb5c7\"" Jan 24 11:42:30.773052 containerd[1625]: time="2026-01-24T11:42:30.772946941Z" level=info msg="StartContainer for \"59672d2115ce67dcb8ee3b08b0d3459c844255fb8bb14dbd511e4b7f35efb5c7\"" Jan 24 11:42:30.777097 containerd[1625]: time="2026-01-24T11:42:30.776121986Z" level=info msg="connecting to shim 59672d2115ce67dcb8ee3b08b0d3459c844255fb8bb14dbd511e4b7f35efb5c7" address="unix:///run/containerd/s/f521e1548f4c377709a04de02e875d100e951e5f99f9cfaf4fb06e5e703c5340" protocol=ttrpc version=3 Jan 24 11:42:30.841069 systemd[1]: Started cri-containerd-534b8923c66f2f4c81f8a584d3fa64023c4f561d2b556be903c5e17f5f031d49.scope - libcontainer container 534b8923c66f2f4c81f8a584d3fa64023c4f561d2b556be903c5e17f5f031d49. Jan 24 11:42:31.133000 audit: BPF prog-id=136 op=LOAD Jan 24 11:42:31.136000 audit: BPF prog-id=137 op=LOAD Jan 24 11:42:31.136000 audit[3125]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001d4238 a2=98 a3=0 items=0 ppid=3114 pid=3125 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.136000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533346238393233633636663266346338316638613538346433666136 Jan 24 11:42:31.136000 audit: BPF prog-id=137 op=UNLOAD Jan 24 11:42:31.136000 audit[3125]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3114 pid=3125 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.136000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533346238393233633636663266346338316638613538346433666136 Jan 24 11:42:31.137000 audit: BPF prog-id=138 op=LOAD Jan 24 11:42:31.137000 audit[3125]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001d4488 a2=98 a3=0 items=0 ppid=3114 pid=3125 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.137000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533346238393233633636663266346338316638613538346433666136 Jan 24 11:42:31.137000 audit: BPF prog-id=139 op=LOAD Jan 24 11:42:31.137000 audit[3125]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001d4218 a2=98 a3=0 items=0 ppid=3114 pid=3125 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.137000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533346238393233633636663266346338316638613538346433666136 Jan 24 11:42:31.137000 audit: BPF prog-id=139 op=UNLOAD Jan 24 11:42:31.137000 audit[3125]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3114 pid=3125 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.137000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533346238393233633636663266346338316638613538346433666136 Jan 24 11:42:31.137000 audit: BPF prog-id=138 op=UNLOAD Jan 24 11:42:31.137000 audit[3125]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3114 pid=3125 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.137000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533346238393233633636663266346338316638613538346433666136 Jan 24 11:42:31.137000 audit: BPF prog-id=140 op=LOAD Jan 24 11:42:31.137000 audit[3125]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001d46e8 a2=98 a3=0 items=0 ppid=3114 pid=3125 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.137000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533346238393233633636663266346338316638613538346433666136 Jan 24 11:42:31.158276 systemd[1]: Started cri-containerd-59672d2115ce67dcb8ee3b08b0d3459c844255fb8bb14dbd511e4b7f35efb5c7.scope - libcontainer container 59672d2115ce67dcb8ee3b08b0d3459c844255fb8bb14dbd511e4b7f35efb5c7. Jan 24 11:42:31.302730 containerd[1625]: time="2026-01-24T11:42:31.301856529Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-65cdcdfd6d-lgqpv,Uid:47852d4e-adf5-4a93-801a-1e88b7d8bc08,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"534b8923c66f2f4c81f8a584d3fa64023c4f561d2b556be903c5e17f5f031d49\"" Jan 24 11:42:31.321716 containerd[1625]: time="2026-01-24T11:42:31.321161585Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Jan 24 11:42:31.343000 audit: BPF prog-id=141 op=LOAD Jan 24 11:42:31.343000 audit[3137]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3068 pid=3137 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.343000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539363732643231313563653637646362386565336230386230643334 Jan 24 11:42:31.343000 audit: BPF prog-id=142 op=LOAD Jan 24 11:42:31.343000 audit[3137]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3068 pid=3137 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.343000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539363732643231313563653637646362386565336230386230643334 Jan 24 11:42:31.343000 audit: BPF prog-id=142 op=UNLOAD Jan 24 11:42:31.343000 audit[3137]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3068 pid=3137 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.343000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539363732643231313563653637646362386565336230386230643334 Jan 24 11:42:31.343000 audit: BPF prog-id=141 op=UNLOAD Jan 24 11:42:31.343000 audit[3137]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3068 pid=3137 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.343000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539363732643231313563653637646362386565336230386230643334 Jan 24 11:42:31.343000 audit: BPF prog-id=143 op=LOAD Jan 24 11:42:31.343000 audit[3137]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3068 pid=3137 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:31.343000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539363732643231313563653637646362386565336230386230643334 Jan 24 11:42:31.392782 containerd[1625]: time="2026-01-24T11:42:31.392547797Z" level=info msg="StartContainer for \"59672d2115ce67dcb8ee3b08b0d3459c844255fb8bb14dbd511e4b7f35efb5c7\" returns successfully" Jan 24 11:42:31.962366 kubelet[2999]: E0124 11:42:31.962135 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:31.996782 kubelet[2999]: I0124 11:42:31.996545 2999 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-4b29f" podStartSLOduration=3.996517487 podStartE2EDuration="3.996517487s" podCreationTimestamp="2026-01-24 11:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:42:31.996370624 +0000 UTC m=+7.795597811" watchObservedRunningTime="2026-01-24 11:42:31.996517487 +0000 UTC m=+7.795744674" Jan 24 11:42:32.396000 audit[3217]: NETFILTER_CFG table=mangle:54 family=2 entries=1 op=nft_register_chain pid=3217 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.396000 audit[3217]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff33369380 a2=0 a3=7fff3336936c items=0 ppid=3157 pid=3217 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.396000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 24 11:42:32.404000 audit[3218]: NETFILTER_CFG table=mangle:55 family=10 entries=1 op=nft_register_chain pid=3218 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.404000 audit[3218]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc27852830 a2=0 a3=7ffc2785281c items=0 ppid=3157 pid=3218 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.404000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 24 11:42:32.408000 audit[3219]: NETFILTER_CFG table=nat:56 family=2 entries=1 op=nft_register_chain pid=3219 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.408000 audit[3219]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc5b155440 a2=0 a3=7ffc5b15542c items=0 ppid=3157 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.408000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 24 11:42:32.417000 audit[3223]: NETFILTER_CFG table=nat:57 family=10 entries=1 op=nft_register_chain pid=3223 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.417000 audit[3223]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff7575d7b0 a2=0 a3=7fff7575d79c items=0 ppid=3157 pid=3223 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.417000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 24 11:42:32.420000 audit[3224]: NETFILTER_CFG table=filter:58 family=2 entries=1 op=nft_register_chain pid=3224 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.420000 audit[3224]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffdec212780 a2=0 a3=7ffdec21276c items=0 ppid=3157 pid=3224 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.420000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 24 11:42:32.424000 audit[3225]: NETFILTER_CFG table=filter:59 family=10 entries=1 op=nft_register_chain pid=3225 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.424000 audit[3225]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff32177220 a2=0 a3=7fff3217720c items=0 ppid=3157 pid=3225 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.424000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 24 11:42:32.518000 audit[3230]: NETFILTER_CFG table=filter:60 family=2 entries=1 op=nft_register_chain pid=3230 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.518000 audit[3230]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffdcdba3fa0 a2=0 a3=7ffdcdba3f8c items=0 ppid=3157 pid=3230 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.518000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 24 11:42:32.523445 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3841112374.mount: Deactivated successfully. Jan 24 11:42:32.528000 audit[3232]: NETFILTER_CFG table=filter:61 family=2 entries=1 op=nft_register_rule pid=3232 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.528000 audit[3232]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffe72647210 a2=0 a3=7ffe726471fc items=0 ppid=3157 pid=3232 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.528000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669636520706F7274616C73002D Jan 24 11:42:32.542000 audit[3235]: NETFILTER_CFG table=filter:62 family=2 entries=1 op=nft_register_rule pid=3235 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.542000 audit[3235]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffe48c666c0 a2=0 a3=7ffe48c666ac items=0 ppid=3157 pid=3235 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.542000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669636520706F7274616C73 Jan 24 11:42:32.546000 audit[3236]: NETFILTER_CFG table=filter:63 family=2 entries=1 op=nft_register_chain pid=3236 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.546000 audit[3236]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffdcacf1ec0 a2=0 a3=7ffdcacf1eac items=0 ppid=3157 pid=3236 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.546000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 24 11:42:32.555000 audit[3238]: NETFILTER_CFG table=filter:64 family=2 entries=1 op=nft_register_rule pid=3238 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.555000 audit[3238]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffd279621c0 a2=0 a3=7ffd279621ac items=0 ppid=3157 pid=3238 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.555000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 24 11:42:32.559000 audit[3239]: NETFILTER_CFG table=filter:65 family=2 entries=1 op=nft_register_chain pid=3239 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.559000 audit[3239]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe933cbd50 a2=0 a3=7ffe933cbd3c items=0 ppid=3157 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.559000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D5345525649434553002D740066696C746572 Jan 24 11:42:32.575000 audit[3241]: NETFILTER_CFG table=filter:66 family=2 entries=1 op=nft_register_rule pid=3241 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.575000 audit[3241]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffd299705a0 a2=0 a3=7ffd2997058c items=0 ppid=3157 pid=3241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.575000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:42:32.587000 audit[3244]: NETFILTER_CFG table=filter:67 family=2 entries=1 op=nft_register_rule pid=3244 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.587000 audit[3244]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffd37883e00 a2=0 a3=7ffd37883dec items=0 ppid=3157 pid=3244 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.587000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:42:32.593000 audit[3245]: NETFILTER_CFG table=filter:68 family=2 entries=1 op=nft_register_chain pid=3245 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.593000 audit[3245]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc901e6ba0 a2=0 a3=7ffc901e6b8c items=0 ppid=3157 pid=3245 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.593000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D464F5257415244002D740066696C746572 Jan 24 11:42:32.602000 audit[3247]: NETFILTER_CFG table=filter:69 family=2 entries=1 op=nft_register_rule pid=3247 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.602000 audit[3247]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffd6ab284e0 a2=0 a3=7ffd6ab284cc items=0 ppid=3157 pid=3247 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.602000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 24 11:42:32.607000 audit[3248]: NETFILTER_CFG table=filter:70 family=2 entries=1 op=nft_register_chain pid=3248 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.607000 audit[3248]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffda7bfa3e0 a2=0 a3=7ffda7bfa3cc items=0 ppid=3157 pid=3248 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.607000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 24 11:42:32.616000 audit[3250]: NETFILTER_CFG table=filter:71 family=2 entries=1 op=nft_register_rule pid=3250 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.616000 audit[3250]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffd9796ea40 a2=0 a3=7ffd9796ea2c items=0 ppid=3157 pid=3250 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.616000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F5859 Jan 24 11:42:32.685000 audit[3253]: NETFILTER_CFG table=filter:72 family=2 entries=1 op=nft_register_rule pid=3253 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.685000 audit[3253]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc793dd920 a2=0 a3=7ffc793dd90c items=0 ppid=3157 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.685000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F58 Jan 24 11:42:32.703000 audit[3256]: NETFILTER_CFG table=filter:73 family=2 entries=1 op=nft_register_rule pid=3256 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.703000 audit[3256]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc8385a010 a2=0 a3=7ffc83859ffc items=0 ppid=3157 pid=3256 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.703000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F Jan 24 11:42:32.707000 audit[3257]: NETFILTER_CFG table=nat:74 family=2 entries=1 op=nft_register_chain pid=3257 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.707000 audit[3257]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffc75141490 a2=0 a3=7ffc7514147c items=0 ppid=3157 pid=3257 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.707000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D5345525649434553002D74006E6174 Jan 24 11:42:32.722000 audit[3259]: NETFILTER_CFG table=nat:75 family=2 entries=1 op=nft_register_rule pid=3259 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.722000 audit[3259]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffd8bbe60c0 a2=0 a3=7ffd8bbe60ac items=0 ppid=3157 pid=3259 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.722000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:42:32.735000 audit[3262]: NETFILTER_CFG table=nat:76 family=2 entries=1 op=nft_register_rule pid=3262 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.735000 audit[3262]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7fff9dd965d0 a2=0 a3=7fff9dd965bc items=0 ppid=3157 pid=3262 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.735000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:42:32.740000 audit[3263]: NETFILTER_CFG table=nat:77 family=2 entries=1 op=nft_register_chain pid=3263 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.740000 audit[3263]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcb8a14a70 a2=0 a3=7ffcb8a14a5c items=0 ppid=3157 pid=3263 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.740000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 24 11:42:32.757000 audit[3265]: NETFILTER_CFG table=nat:78 family=2 entries=1 op=nft_register_rule pid=3265 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:42:32.757000 audit[3265]: SYSCALL arch=c000003e syscall=46 success=yes exit=532 a0=3 a1=7ffe84975f00 a2=0 a3=7ffe84975eec items=0 ppid=3157 pid=3265 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.757000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 24 11:42:32.814000 audit[3271]: NETFILTER_CFG table=filter:79 family=2 entries=8 op=nft_register_rule pid=3271 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:32.814000 audit[3271]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffc3a991780 a2=0 a3=7ffc3a99176c items=0 ppid=3157 pid=3271 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.814000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:32.832000 audit[3271]: NETFILTER_CFG table=nat:80 family=2 entries=14 op=nft_register_chain pid=3271 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:32.832000 audit[3271]: SYSCALL arch=c000003e syscall=46 success=yes exit=5508 a0=3 a1=7ffc3a991780 a2=0 a3=7ffc3a99176c items=0 ppid=3157 pid=3271 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.832000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:32.837000 audit[3281]: NETFILTER_CFG table=filter:81 family=10 entries=1 op=nft_register_chain pid=3281 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.837000 audit[3281]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffed8f55aa0 a2=0 a3=7ffed8f55a8c items=0 ppid=3157 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.837000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 24 11:42:32.844000 audit[3283]: NETFILTER_CFG table=filter:82 family=10 entries=2 op=nft_register_chain pid=3283 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.844000 audit[3283]: SYSCALL arch=c000003e syscall=46 success=yes exit=836 a0=3 a1=7ffe77c6a370 a2=0 a3=7ffe77c6a35c items=0 ppid=3157 pid=3283 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.844000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669636520706F7274616C73 Jan 24 11:42:32.862000 audit[3286]: NETFILTER_CFG table=filter:83 family=10 entries=1 op=nft_register_rule pid=3286 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.862000 audit[3286]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffea4b6dc00 a2=0 a3=7ffea4b6dbec items=0 ppid=3157 pid=3286 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.862000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669636520706F7274616C Jan 24 11:42:32.869000 audit[3287]: NETFILTER_CFG table=filter:84 family=10 entries=1 op=nft_register_chain pid=3287 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.869000 audit[3287]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe9a3f6070 a2=0 a3=7ffe9a3f605c items=0 ppid=3157 pid=3287 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.869000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 24 11:42:32.887000 audit[3289]: NETFILTER_CFG table=filter:85 family=10 entries=1 op=nft_register_rule pid=3289 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.887000 audit[3289]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffd8236fb50 a2=0 a3=7ffd8236fb3c items=0 ppid=3157 pid=3289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.887000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 24 11:42:32.896000 audit[3290]: NETFILTER_CFG table=filter:86 family=10 entries=1 op=nft_register_chain pid=3290 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.896000 audit[3290]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc17605ea0 a2=0 a3=7ffc17605e8c items=0 ppid=3157 pid=3290 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.896000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D5345525649434553002D740066696C746572 Jan 24 11:42:32.916000 audit[3292]: NETFILTER_CFG table=filter:87 family=10 entries=1 op=nft_register_rule pid=3292 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.916000 audit[3292]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffc89ea0880 a2=0 a3=7ffc89ea086c items=0 ppid=3157 pid=3292 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.916000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:42:32.931000 audit[3295]: NETFILTER_CFG table=filter:88 family=10 entries=2 op=nft_register_chain pid=3295 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.931000 audit[3295]: SYSCALL arch=c000003e syscall=46 success=yes exit=828 a0=3 a1=7ffe70949fc0 a2=0 a3=7ffe70949fac items=0 ppid=3157 pid=3295 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.931000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:42:32.935000 audit[3296]: NETFILTER_CFG table=filter:89 family=10 entries=1 op=nft_register_chain pid=3296 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.935000 audit[3296]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd51f8b2a0 a2=0 a3=7ffd51f8b28c items=0 ppid=3157 pid=3296 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.935000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D464F5257415244002D740066696C746572 Jan 24 11:42:32.944000 audit[3298]: NETFILTER_CFG table=filter:90 family=10 entries=1 op=nft_register_rule pid=3298 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.944000 audit[3298]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffdb94740d0 a2=0 a3=7ffdb94740bc items=0 ppid=3157 pid=3298 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.944000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 24 11:42:32.950000 audit[3299]: NETFILTER_CFG table=filter:91 family=10 entries=1 op=nft_register_chain pid=3299 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.950000 audit[3299]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc206a67e0 a2=0 a3=7ffc206a67cc items=0 ppid=3157 pid=3299 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.950000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 24 11:42:32.963000 audit[3301]: NETFILTER_CFG table=filter:92 family=10 entries=1 op=nft_register_rule pid=3301 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.963000 audit[3301]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7fff25b57350 a2=0 a3=7fff25b5733c items=0 ppid=3157 pid=3301 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.963000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F58 Jan 24 11:42:32.971050 kubelet[2999]: E0124 11:42:32.970417 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:32.979000 audit[3304]: NETFILTER_CFG table=filter:93 family=10 entries=1 op=nft_register_rule pid=3304 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.979000 audit[3304]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7fff44db6ab0 a2=0 a3=7fff44db6a9c items=0 ppid=3157 pid=3304 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.979000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F Jan 24 11:42:32.993000 audit[3307]: NETFILTER_CFG table=filter:94 family=10 entries=1 op=nft_register_rule pid=3307 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.993000 audit[3307]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffd7e478e60 a2=0 a3=7ffd7e478e4c items=0 ppid=3157 pid=3307 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.993000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D5052 Jan 24 11:42:32.998000 audit[3308]: NETFILTER_CFG table=nat:95 family=10 entries=1 op=nft_register_chain pid=3308 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:32.998000 audit[3308]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7fff517a8c90 a2=0 a3=7fff517a8c7c items=0 ppid=3157 pid=3308 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:32.998000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D5345525649434553002D74006E6174 Jan 24 11:42:33.007000 audit[3310]: NETFILTER_CFG table=nat:96 family=10 entries=1 op=nft_register_rule pid=3310 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:33.007000 audit[3310]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffe476f5cc0 a2=0 a3=7ffe476f5cac items=0 ppid=3157 pid=3310 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:33.007000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:42:33.018000 audit[3313]: NETFILTER_CFG table=nat:97 family=10 entries=1 op=nft_register_rule pid=3313 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:33.018000 audit[3313]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7fff4a498d60 a2=0 a3=7fff4a498d4c items=0 ppid=3157 pid=3313 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:33.018000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:42:33.030000 audit[3314]: NETFILTER_CFG table=nat:98 family=10 entries=1 op=nft_register_chain pid=3314 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:33.030000 audit[3314]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe8ad7d1b0 a2=0 a3=7ffe8ad7d19c items=0 ppid=3157 pid=3314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:33.030000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 24 11:42:33.041000 audit[3316]: NETFILTER_CFG table=nat:99 family=10 entries=2 op=nft_register_chain pid=3316 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:33.041000 audit[3316]: SYSCALL arch=c000003e syscall=46 success=yes exit=612 a0=3 a1=7ffd13889f70 a2=0 a3=7ffd13889f5c items=0 ppid=3157 pid=3316 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:33.041000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 24 11:42:33.045000 audit[3317]: NETFILTER_CFG table=filter:100 family=10 entries=1 op=nft_register_chain pid=3317 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:33.045000 audit[3317]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd5af797e0 a2=0 a3=7ffd5af797cc items=0 ppid=3157 pid=3317 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:33.045000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 24 11:42:33.053000 audit[3319]: NETFILTER_CFG table=filter:101 family=10 entries=1 op=nft_register_rule pid=3319 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:33.053000 audit[3319]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffcbe73f700 a2=0 a3=7ffcbe73f6ec items=0 ppid=3157 pid=3319 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:33.053000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 24 11:42:33.067000 audit[3322]: NETFILTER_CFG table=filter:102 family=10 entries=1 op=nft_register_rule pid=3322 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:42:33.067000 audit[3322]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7fff4e645490 a2=0 a3=7fff4e64547c items=0 ppid=3157 pid=3322 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:33.067000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 24 11:42:33.081000 audit[3324]: NETFILTER_CFG table=filter:103 family=10 entries=3 op=nft_register_rule pid=3324 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 24 11:42:33.081000 audit[3324]: SYSCALL arch=c000003e syscall=46 success=yes exit=2088 a0=3 a1=7ffdf3a817c0 a2=0 a3=7ffdf3a817ac items=0 ppid=3157 pid=3324 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:33.081000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:33.082000 audit[3324]: NETFILTER_CFG table=nat:104 family=10 entries=7 op=nft_register_chain pid=3324 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 24 11:42:33.082000 audit[3324]: SYSCALL arch=c000003e syscall=46 success=yes exit=2056 a0=3 a1=7ffdf3a817c0 a2=0 a3=7ffdf3a817ac items=0 ppid=3157 pid=3324 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:33.082000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:34.096921 kubelet[2999]: E0124 11:42:34.096434 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:34.479778 containerd[1625]: time="2026-01-24T11:42:34.479440731Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:42:34.484159 containerd[1625]: time="2026-01-24T11:42:34.483548880Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=23558205" Jan 24 11:42:34.487051 containerd[1625]: time="2026-01-24T11:42:34.486936826Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:42:34.535576 containerd[1625]: time="2026-01-24T11:42:34.534805862Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:42:34.545978 containerd[1625]: time="2026-01-24T11:42:34.545165948Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 3.223652247s" Jan 24 11:42:34.545978 containerd[1625]: time="2026-01-24T11:42:34.545370598Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Jan 24 11:42:34.568805 containerd[1625]: time="2026-01-24T11:42:34.568744512Z" level=info msg="CreateContainer within sandbox \"534b8923c66f2f4c81f8a584d3fa64023c4f561d2b556be903c5e17f5f031d49\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jan 24 11:42:34.635401 containerd[1625]: time="2026-01-24T11:42:34.635064307Z" level=info msg="Container 8198b48daf6507534ae3df871332f6873fcea950cfac149328e6eed022196eea: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:42:34.666257 containerd[1625]: time="2026-01-24T11:42:34.666151695Z" level=info msg="CreateContainer within sandbox \"534b8923c66f2f4c81f8a584d3fa64023c4f561d2b556be903c5e17f5f031d49\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"8198b48daf6507534ae3df871332f6873fcea950cfac149328e6eed022196eea\"" Jan 24 11:42:34.667749 containerd[1625]: time="2026-01-24T11:42:34.667675540Z" level=info msg="StartContainer for \"8198b48daf6507534ae3df871332f6873fcea950cfac149328e6eed022196eea\"" Jan 24 11:42:34.671919 containerd[1625]: time="2026-01-24T11:42:34.671729492Z" level=info msg="connecting to shim 8198b48daf6507534ae3df871332f6873fcea950cfac149328e6eed022196eea" address="unix:///run/containerd/s/2bcdf48c2c7ad57d1d79c02329d7eed9a4279291750057beb3562ceaabaed610" protocol=ttrpc version=3 Jan 24 11:42:34.787805 systemd[1]: Started cri-containerd-8198b48daf6507534ae3df871332f6873fcea950cfac149328e6eed022196eea.scope - libcontainer container 8198b48daf6507534ae3df871332f6873fcea950cfac149328e6eed022196eea. Jan 24 11:42:34.916000 audit: BPF prog-id=144 op=LOAD Jan 24 11:42:34.922000 audit: BPF prog-id=145 op=LOAD Jan 24 11:42:34.922000 audit[3326]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=3114 pid=3326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:34.922000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831393862343864616636353037353334616533646638373133333266 Jan 24 11:42:34.922000 audit: BPF prog-id=145 op=UNLOAD Jan 24 11:42:34.922000 audit[3326]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3114 pid=3326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:34.922000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831393862343864616636353037353334616533646638373133333266 Jan 24 11:42:34.926000 audit: BPF prog-id=146 op=LOAD Jan 24 11:42:34.926000 audit[3326]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3114 pid=3326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:34.926000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831393862343864616636353037353334616533646638373133333266 Jan 24 11:42:34.927000 audit: BPF prog-id=147 op=LOAD Jan 24 11:42:34.927000 audit[3326]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=3114 pid=3326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:34.927000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831393862343864616636353037353334616533646638373133333266 Jan 24 11:42:34.927000 audit: BPF prog-id=147 op=UNLOAD Jan 24 11:42:34.927000 audit[3326]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3114 pid=3326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:34.927000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831393862343864616636353037353334616533646638373133333266 Jan 24 11:42:34.927000 audit: BPF prog-id=146 op=UNLOAD Jan 24 11:42:34.927000 audit[3326]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3114 pid=3326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:34.927000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831393862343864616636353037353334616533646638373133333266 Jan 24 11:42:34.928000 audit: BPF prog-id=148 op=LOAD Jan 24 11:42:34.928000 audit[3326]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=3114 pid=3326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:34.928000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831393862343864616636353037353334616533646638373133333266 Jan 24 11:42:35.124554 kubelet[2999]: E0124 11:42:35.122323 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:35.132145 containerd[1625]: time="2026-01-24T11:42:35.131349221Z" level=info msg="StartContainer for \"8198b48daf6507534ae3df871332f6873fcea950cfac149328e6eed022196eea\" returns successfully" Jan 24 11:42:36.147498 kubelet[2999]: E0124 11:42:36.146752 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:36.225940 kubelet[2999]: I0124 11:42:36.225582 2999 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-65cdcdfd6d-lgqpv" podStartSLOduration=3.988967765 podStartE2EDuration="7.225566705s" podCreationTimestamp="2026-01-24 11:42:29 +0000 UTC" firstStartedPulling="2026-01-24 11:42:31.319096416 +0000 UTC m=+7.118323603" lastFinishedPulling="2026-01-24 11:42:34.555695356 +0000 UTC m=+10.354922543" observedRunningTime="2026-01-24 11:42:36.224557246 +0000 UTC m=+12.023784453" watchObservedRunningTime="2026-01-24 11:42:36.225566705 +0000 UTC m=+12.024793892" Jan 24 11:42:44.332238 kernel: kauditd_printk_skb: 224 callbacks suppressed Jan 24 11:42:44.332441 kernel: audit: type=1106 audit(1769254964.311:526): pid=1854 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:42:44.311000 audit[1854]: USER_END pid=1854 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:42:44.312893 sudo[1854]: pam_unix(sudo:session): session closed for user root Jan 24 11:42:44.317000 audit[1854]: CRED_DISP pid=1854 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:42:44.346384 kernel: audit: type=1104 audit(1769254964.317:527): pid=1854 uid=500 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:42:44.350433 sshd[1853]: Connection closed by 10.0.0.1 port 51844 Jan 24 11:42:44.356795 sshd-session[1849]: pam_unix(sshd:session): session closed for user core Jan 24 11:42:44.361000 audit[1849]: USER_END pid=1849 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:42:44.370290 systemd[1]: sshd@8-10.0.0.26:22-10.0.0.1:51844.service: Deactivated successfully. Jan 24 11:42:44.361000 audit[1849]: CRED_DISP pid=1849 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:42:44.377035 systemd[1]: session-10.scope: Deactivated successfully. Jan 24 11:42:44.378743 systemd[1]: session-10.scope: Consumed 20.975s CPU time, 227M memory peak. Jan 24 11:42:44.385592 systemd-logind[1579]: Session 10 logged out. Waiting for processes to exit. Jan 24 11:42:44.391956 systemd-logind[1579]: Removed session 10. Jan 24 11:42:44.408170 kernel: audit: type=1106 audit(1769254964.361:528): pid=1849 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:42:44.408420 kernel: audit: type=1104 audit(1769254964.361:529): pid=1849 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:42:44.368000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.26:22-10.0.0.1:51844 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:42:44.426451 kernel: audit: type=1131 audit(1769254964.368:530): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.26:22-10.0.0.1:51844 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:42:45.115000 audit[3420]: NETFILTER_CFG table=filter:105 family=2 entries=15 op=nft_register_rule pid=3420 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:45.115000 audit[3420]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7fff07999b50 a2=0 a3=7fff07999b3c items=0 ppid=3157 pid=3420 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:45.149917 kernel: audit: type=1325 audit(1769254965.115:531): table=filter:105 family=2 entries=15 op=nft_register_rule pid=3420 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:45.150094 kernel: audit: type=1300 audit(1769254965.115:531): arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7fff07999b50 a2=0 a3=7fff07999b3c items=0 ppid=3157 pid=3420 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:45.115000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:45.167324 kernel: audit: type=1327 audit(1769254965.115:531): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:45.153000 audit[3420]: NETFILTER_CFG table=nat:106 family=2 entries=12 op=nft_register_rule pid=3420 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:45.187280 kernel: audit: type=1325 audit(1769254965.153:532): table=nat:106 family=2 entries=12 op=nft_register_rule pid=3420 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:45.153000 audit[3420]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff07999b50 a2=0 a3=0 items=0 ppid=3157 pid=3420 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:45.153000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:45.211366 kernel: audit: type=1300 audit(1769254965.153:532): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff07999b50 a2=0 a3=0 items=0 ppid=3157 pid=3420 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:46.263000 audit[3423]: NETFILTER_CFG table=filter:107 family=2 entries=16 op=nft_register_rule pid=3423 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:46.263000 audit[3423]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7fff076e1550 a2=0 a3=7fff076e153c items=0 ppid=3157 pid=3423 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:46.263000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:46.269000 audit[3423]: NETFILTER_CFG table=nat:108 family=2 entries=12 op=nft_register_rule pid=3423 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:46.269000 audit[3423]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff076e1550 a2=0 a3=0 items=0 ppid=3157 pid=3423 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:46.269000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:49.509092 kernel: kauditd_printk_skb: 7 callbacks suppressed Jan 24 11:42:49.509375 kernel: audit: type=1325 audit(1769254969.491:535): table=filter:109 family=2 entries=17 op=nft_register_rule pid=3425 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:49.491000 audit[3425]: NETFILTER_CFG table=filter:109 family=2 entries=17 op=nft_register_rule pid=3425 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:49.491000 audit[3425]: SYSCALL arch=c000003e syscall=46 success=yes exit=6736 a0=3 a1=7ffd51dd4590 a2=0 a3=7ffd51dd457c items=0 ppid=3157 pid=3425 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:49.549486 kernel: audit: type=1300 audit(1769254969.491:535): arch=c000003e syscall=46 success=yes exit=6736 a0=3 a1=7ffd51dd4590 a2=0 a3=7ffd51dd457c items=0 ppid=3157 pid=3425 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:49.550012 kernel: audit: type=1327 audit(1769254969.491:535): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:49.491000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:49.562355 kernel: audit: type=1325 audit(1769254969.542:536): table=nat:110 family=2 entries=12 op=nft_register_rule pid=3425 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:49.542000 audit[3425]: NETFILTER_CFG table=nat:110 family=2 entries=12 op=nft_register_rule pid=3425 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:49.542000 audit[3425]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd51dd4590 a2=0 a3=0 items=0 ppid=3157 pid=3425 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:49.589644 kernel: audit: type=1300 audit(1769254969.542:536): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd51dd4590 a2=0 a3=0 items=0 ppid=3157 pid=3425 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:49.589789 kernel: audit: type=1327 audit(1769254969.542:536): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:49.542000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:50.611000 audit[3427]: NETFILTER_CFG table=filter:111 family=2 entries=19 op=nft_register_rule pid=3427 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:50.630292 kernel: audit: type=1325 audit(1769254970.611:537): table=filter:111 family=2 entries=19 op=nft_register_rule pid=3427 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:50.630770 kernel: audit: type=1300 audit(1769254970.611:537): arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffce1d57160 a2=0 a3=7ffce1d5714c items=0 ppid=3157 pid=3427 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:50.611000 audit[3427]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffce1d57160 a2=0 a3=7ffce1d5714c items=0 ppid=3157 pid=3427 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:50.611000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:50.661000 audit[3427]: NETFILTER_CFG table=nat:112 family=2 entries=12 op=nft_register_rule pid=3427 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:50.682348 kernel: audit: type=1327 audit(1769254970.611:537): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:50.682493 kernel: audit: type=1325 audit(1769254970.661:538): table=nat:112 family=2 entries=12 op=nft_register_rule pid=3427 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:50.661000 audit[3427]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffce1d57160 a2=0 a3=0 items=0 ppid=3157 pid=3427 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:50.661000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:52.665000 audit[3430]: NETFILTER_CFG table=filter:113 family=2 entries=21 op=nft_register_rule pid=3430 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:52.665000 audit[3430]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffd38de55d0 a2=0 a3=7ffd38de55bc items=0 ppid=3157 pid=3430 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:52.665000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:52.689000 audit[3430]: NETFILTER_CFG table=nat:114 family=2 entries=12 op=nft_register_rule pid=3430 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:52.689000 audit[3430]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd38de55d0 a2=0 a3=0 items=0 ppid=3157 pid=3430 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:52.689000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:52.827546 systemd[1]: Created slice kubepods-besteffort-podadf2c65c_497b_4e5d_b8ba_e6f639ae408e.slice - libcontainer container kubepods-besteffort-podadf2c65c_497b_4e5d_b8ba_e6f639ae408e.slice. Jan 24 11:42:52.990546 kubelet[2999]: I0124 11:42:52.989670 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/adf2c65c-497b-4e5d-b8ba-e6f639ae408e-typha-certs\") pod \"calico-typha-8dcb8f48d-jl7pj\" (UID: \"adf2c65c-497b-4e5d-b8ba-e6f639ae408e\") " pod="calico-system/calico-typha-8dcb8f48d-jl7pj" Jan 24 11:42:52.990546 kubelet[2999]: I0124 11:42:52.989892 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z2tx\" (UniqueName: \"kubernetes.io/projected/adf2c65c-497b-4e5d-b8ba-e6f639ae408e-kube-api-access-2z2tx\") pod \"calico-typha-8dcb8f48d-jl7pj\" (UID: \"adf2c65c-497b-4e5d-b8ba-e6f639ae408e\") " pod="calico-system/calico-typha-8dcb8f48d-jl7pj" Jan 24 11:42:52.990546 kubelet[2999]: I0124 11:42:52.989922 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/adf2c65c-497b-4e5d-b8ba-e6f639ae408e-tigera-ca-bundle\") pod \"calico-typha-8dcb8f48d-jl7pj\" (UID: \"adf2c65c-497b-4e5d-b8ba-e6f639ae408e\") " pod="calico-system/calico-typha-8dcb8f48d-jl7pj" Jan 24 11:42:53.050074 systemd[1]: Created slice kubepods-besteffort-pod49b6127d_07b4_4854_8332_f6cab36ded5e.slice - libcontainer container kubepods-besteffort-pod49b6127d_07b4_4854_8332_f6cab36ded5e.slice. Jan 24 11:42:53.092143 kubelet[2999]: I0124 11:42:53.091774 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49b6127d-07b4-4854-8332-f6cab36ded5e-tigera-ca-bundle\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.092143 kubelet[2999]: I0124 11:42:53.091942 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/49b6127d-07b4-4854-8332-f6cab36ded5e-var-lib-calico\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.092143 kubelet[2999]: I0124 11:42:53.091971 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/49b6127d-07b4-4854-8332-f6cab36ded5e-xtables-lock\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.092143 kubelet[2999]: I0124 11:42:53.091996 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/49b6127d-07b4-4854-8332-f6cab36ded5e-flexvol-driver-host\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.092143 kubelet[2999]: I0124 11:42:53.092026 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/49b6127d-07b4-4854-8332-f6cab36ded5e-lib-modules\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.094588 kubelet[2999]: I0124 11:42:53.092063 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/49b6127d-07b4-4854-8332-f6cab36ded5e-node-certs\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.094588 kubelet[2999]: I0124 11:42:53.092085 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/49b6127d-07b4-4854-8332-f6cab36ded5e-policysync\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.094588 kubelet[2999]: I0124 11:42:53.092106 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/49b6127d-07b4-4854-8332-f6cab36ded5e-var-run-calico\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.094588 kubelet[2999]: I0124 11:42:53.092130 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ckw9\" (UniqueName: \"kubernetes.io/projected/49b6127d-07b4-4854-8332-f6cab36ded5e-kube-api-access-8ckw9\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.094588 kubelet[2999]: I0124 11:42:53.092165 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/49b6127d-07b4-4854-8332-f6cab36ded5e-cni-bin-dir\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.094897 kubelet[2999]: I0124 11:42:53.092349 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/49b6127d-07b4-4854-8332-f6cab36ded5e-cni-log-dir\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.094897 kubelet[2999]: I0124 11:42:53.092383 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/49b6127d-07b4-4854-8332-f6cab36ded5e-cni-net-dir\") pod \"calico-node-knsxr\" (UID: \"49b6127d-07b4-4854-8332-f6cab36ded5e\") " pod="calico-system/calico-node-knsxr" Jan 24 11:42:53.150795 systemd[1682]: Created slice background.slice - User Background Tasks Slice. Jan 24 11:42:53.157325 systemd[1682]: Starting systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories... Jan 24 11:42:53.207563 kubelet[2999]: E0124 11:42:53.206973 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:42:53.238643 systemd[1682]: Finished systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories. Jan 24 11:42:53.247378 kubelet[2999]: E0124 11:42:53.246940 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.247378 kubelet[2999]: W0124 11:42:53.246974 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.247378 kubelet[2999]: E0124 11:42:53.247175 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.263353 kubelet[2999]: E0124 11:42:53.263128 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.263983 kubelet[2999]: W0124 11:42:53.263870 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.263983 kubelet[2999]: E0124 11:42:53.263954 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.295776 kubelet[2999]: E0124 11:42:53.295309 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.295776 kubelet[2999]: W0124 11:42:53.295347 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.295776 kubelet[2999]: E0124 11:42:53.295378 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.297043 kubelet[2999]: E0124 11:42:53.297020 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.297157 kubelet[2999]: W0124 11:42:53.297139 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.297612 kubelet[2999]: E0124 11:42:53.297414 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.302933 kubelet[2999]: E0124 11:42:53.302911 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.303050 kubelet[2999]: W0124 11:42:53.303029 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.303159 kubelet[2999]: E0124 11:42:53.303137 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.304603 kubelet[2999]: E0124 11:42:53.304331 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.304603 kubelet[2999]: W0124 11:42:53.304394 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.304603 kubelet[2999]: E0124 11:42:53.304415 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.305034 kubelet[2999]: E0124 11:42:53.304917 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.305034 kubelet[2999]: W0124 11:42:53.304989 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.305034 kubelet[2999]: E0124 11:42:53.305005 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.305752 kubelet[2999]: E0124 11:42:53.305547 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.305752 kubelet[2999]: W0124 11:42:53.305615 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.305752 kubelet[2999]: E0124 11:42:53.305632 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.306494 kubelet[2999]: E0124 11:42:53.306145 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.306494 kubelet[2999]: W0124 11:42:53.306160 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.306494 kubelet[2999]: E0124 11:42:53.306176 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.307520 kubelet[2999]: E0124 11:42:53.307361 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.307520 kubelet[2999]: W0124 11:42:53.307381 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.307520 kubelet[2999]: E0124 11:42:53.307395 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.310027 kubelet[2999]: E0124 11:42:53.308031 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.310027 kubelet[2999]: W0124 11:42:53.308045 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.310027 kubelet[2999]: E0124 11:42:53.308059 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.310027 kubelet[2999]: E0124 11:42:53.309159 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.310027 kubelet[2999]: W0124 11:42:53.309172 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.310027 kubelet[2999]: E0124 11:42:53.309334 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.310027 kubelet[2999]: E0124 11:42:53.309857 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.310027 kubelet[2999]: W0124 11:42:53.309872 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.310027 kubelet[2999]: E0124 11:42:53.309886 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.310860 kubelet[2999]: E0124 11:42:53.310366 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.310860 kubelet[2999]: W0124 11:42:53.310381 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.310860 kubelet[2999]: E0124 11:42:53.310394 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.310860 kubelet[2999]: E0124 11:42:53.310794 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.310860 kubelet[2999]: W0124 11:42:53.310868 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.311130 kubelet[2999]: E0124 11:42:53.310882 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.311646 kubelet[2999]: E0124 11:42:53.311391 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.311646 kubelet[2999]: W0124 11:42:53.311407 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.311646 kubelet[2999]: E0124 11:42:53.311421 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.311646 kubelet[2999]: E0124 11:42:53.311792 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.311646 kubelet[2999]: W0124 11:42:53.311865 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.311646 kubelet[2999]: E0124 11:42:53.311878 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.311646 kubelet[2999]: E0124 11:42:53.312135 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.311646 kubelet[2999]: W0124 11:42:53.312148 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.311646 kubelet[2999]: E0124 11:42:53.312160 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.316117 kubelet[2999]: E0124 11:42:53.313470 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.316117 kubelet[2999]: W0124 11:42:53.313481 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.316117 kubelet[2999]: E0124 11:42:53.313493 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.316117 kubelet[2999]: E0124 11:42:53.313728 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.316117 kubelet[2999]: W0124 11:42:53.313737 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.316117 kubelet[2999]: E0124 11:42:53.313746 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.316117 kubelet[2999]: E0124 11:42:53.314041 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.316117 kubelet[2999]: W0124 11:42:53.314051 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.316117 kubelet[2999]: E0124 11:42:53.314072 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.316117 kubelet[2999]: E0124 11:42:53.314548 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.316770 kubelet[2999]: W0124 11:42:53.314558 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.316770 kubelet[2999]: E0124 11:42:53.314568 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.316770 kubelet[2999]: E0124 11:42:53.316335 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.316770 kubelet[2999]: W0124 11:42:53.316349 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.316770 kubelet[2999]: E0124 11:42:53.316361 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.318479 kubelet[2999]: I0124 11:42:53.317791 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd-socket-dir\") pod \"csi-node-driver-j5qxv\" (UID: \"9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd\") " pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:42:53.321366 kubelet[2999]: E0124 11:42:53.320501 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.321366 kubelet[2999]: W0124 11:42:53.320522 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.321366 kubelet[2999]: E0124 11:42:53.320536 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.321366 kubelet[2999]: I0124 11:42:53.321174 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4vwl\" (UniqueName: \"kubernetes.io/projected/9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd-kube-api-access-x4vwl\") pod \"csi-node-driver-j5qxv\" (UID: \"9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd\") " pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:42:53.323648 kubelet[2999]: E0124 11:42:53.323532 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.323648 kubelet[2999]: W0124 11:42:53.323605 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.323648 kubelet[2999]: E0124 11:42:53.323620 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.328038 kubelet[2999]: E0124 11:42:53.327753 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.328038 kubelet[2999]: W0124 11:42:53.327878 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.328038 kubelet[2999]: E0124 11:42:53.327893 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.330025 kubelet[2999]: E0124 11:42:53.329885 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.330025 kubelet[2999]: W0124 11:42:53.329913 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.330025 kubelet[2999]: E0124 11:42:53.329930 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.330520 kubelet[2999]: I0124 11:42:53.330481 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd-registration-dir\") pod \"csi-node-driver-j5qxv\" (UID: \"9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd\") " pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:42:53.331550 kubelet[2999]: E0124 11:42:53.331381 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.333003 kubelet[2999]: W0124 11:42:53.331613 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.333003 kubelet[2999]: E0124 11:42:53.332505 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.338012 kubelet[2999]: E0124 11:42:53.337683 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.338012 kubelet[2999]: W0124 11:42:53.337770 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.338012 kubelet[2999]: E0124 11:42:53.337795 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.340074 kubelet[2999]: E0124 11:42:53.339332 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.340074 kubelet[2999]: W0124 11:42:53.339347 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.340074 kubelet[2999]: E0124 11:42:53.339363 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.341069 kubelet[2999]: I0124 11:42:53.340958 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd-kubelet-dir\") pod \"csi-node-driver-j5qxv\" (UID: \"9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd\") " pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:42:53.342690 kubelet[2999]: E0124 11:42:53.342571 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.342690 kubelet[2999]: W0124 11:42:53.342587 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.342690 kubelet[2999]: E0124 11:42:53.342605 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.345509 kubelet[2999]: E0124 11:42:53.345147 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.345509 kubelet[2999]: W0124 11:42:53.345399 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.345509 kubelet[2999]: E0124 11:42:53.345422 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.348137 kubelet[2999]: E0124 11:42:53.347942 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.348137 kubelet[2999]: W0124 11:42:53.347973 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.348137 kubelet[2999]: E0124 11:42:53.347991 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.348137 kubelet[2999]: I0124 11:42:53.348031 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd-varrun\") pod \"csi-node-driver-j5qxv\" (UID: \"9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd\") " pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:42:53.352455 kubelet[2999]: E0124 11:42:53.351874 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.352455 kubelet[2999]: W0124 11:42:53.351910 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.352455 kubelet[2999]: E0124 11:42:53.351935 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.353672 kubelet[2999]: E0124 11:42:53.353546 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.353672 kubelet[2999]: W0124 11:42:53.353612 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.353672 kubelet[2999]: E0124 11:42:53.353627 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.357739 kubelet[2999]: E0124 11:42:53.357429 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.357739 kubelet[2999]: W0124 11:42:53.357514 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.357739 kubelet[2999]: E0124 11:42:53.357531 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.358872 kubelet[2999]: E0124 11:42:53.358580 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.358872 kubelet[2999]: W0124 11:42:53.358598 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.358872 kubelet[2999]: E0124 11:42:53.358611 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.367153 kubelet[2999]: E0124 11:42:53.366504 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:53.368649 containerd[1625]: time="2026-01-24T11:42:53.368527255Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-knsxr,Uid:49b6127d-07b4-4854-8332-f6cab36ded5e,Namespace:calico-system,Attempt:0,}" Jan 24 11:42:53.448165 kubelet[2999]: E0124 11:42:53.447893 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:53.450482 containerd[1625]: time="2026-01-24T11:42:53.450433726Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-8dcb8f48d-jl7pj,Uid:adf2c65c-497b-4e5d-b8ba-e6f639ae408e,Namespace:calico-system,Attempt:0,}" Jan 24 11:42:53.452389 kubelet[2999]: E0124 11:42:53.452366 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.454133 kubelet[2999]: W0124 11:42:53.453154 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.454133 kubelet[2999]: E0124 11:42:53.453363 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.455400 kubelet[2999]: E0124 11:42:53.455384 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.455633 kubelet[2999]: W0124 11:42:53.455532 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.455782 kubelet[2999]: E0124 11:42:53.455697 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.461335 kubelet[2999]: E0124 11:42:53.460966 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.461335 kubelet[2999]: W0124 11:42:53.460988 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.461335 kubelet[2999]: E0124 11:42:53.461013 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.462088 kubelet[2999]: E0124 11:42:53.462032 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.462088 kubelet[2999]: W0124 11:42:53.462052 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.462088 kubelet[2999]: E0124 11:42:53.462070 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.464123 kubelet[2999]: E0124 11:42:53.463728 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.464123 kubelet[2999]: W0124 11:42:53.463873 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.464123 kubelet[2999]: E0124 11:42:53.463897 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.466344 kubelet[2999]: E0124 11:42:53.466044 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.466344 kubelet[2999]: W0124 11:42:53.466128 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.466344 kubelet[2999]: E0124 11:42:53.466148 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.466906 kubelet[2999]: E0124 11:42:53.466764 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.466953 kubelet[2999]: W0124 11:42:53.466907 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.466953 kubelet[2999]: E0124 11:42:53.466924 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.468742 kubelet[2999]: E0124 11:42:53.467996 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.468742 kubelet[2999]: W0124 11:42:53.468073 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.468742 kubelet[2999]: E0124 11:42:53.468089 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.470646 kubelet[2999]: E0124 11:42:53.470546 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.470646 kubelet[2999]: W0124 11:42:53.470632 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.470646 kubelet[2999]: E0124 11:42:53.470654 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.472779 kubelet[2999]: E0124 11:42:53.472631 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.472779 kubelet[2999]: W0124 11:42:53.472717 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.472779 kubelet[2999]: E0124 11:42:53.472735 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.475733 kubelet[2999]: E0124 11:42:53.475655 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.475733 kubelet[2999]: W0124 11:42:53.475694 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.475733 kubelet[2999]: E0124 11:42:53.475711 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.478612 kubelet[2999]: E0124 11:42:53.478552 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.478612 kubelet[2999]: W0124 11:42:53.478575 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.478612 kubelet[2999]: E0124 11:42:53.478592 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.482573 kubelet[2999]: E0124 11:42:53.481518 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.482573 kubelet[2999]: W0124 11:42:53.481535 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.482573 kubelet[2999]: E0124 11:42:53.481552 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.483590 kubelet[2999]: E0124 11:42:53.482929 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.483590 kubelet[2999]: W0124 11:42:53.482946 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.483590 kubelet[2999]: E0124 11:42:53.482960 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.486489 kubelet[2999]: E0124 11:42:53.486471 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.486572 kubelet[2999]: W0124 11:42:53.486553 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.486686 kubelet[2999]: E0124 11:42:53.486666 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.490099 kubelet[2999]: E0124 11:42:53.490079 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.490362 kubelet[2999]: W0124 11:42:53.490165 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.490446 kubelet[2999]: E0124 11:42:53.490430 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.494167 kubelet[2999]: E0124 11:42:53.493474 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.494167 kubelet[2999]: W0124 11:42:53.493490 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.494167 kubelet[2999]: E0124 11:42:53.493505 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.499382 kubelet[2999]: E0124 11:42:53.497170 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.499503 kubelet[2999]: W0124 11:42:53.499483 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.501527 kubelet[2999]: E0124 11:42:53.500753 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.508658 kubelet[2999]: E0124 11:42:53.508590 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.508658 kubelet[2999]: W0124 11:42:53.508611 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.508658 kubelet[2999]: E0124 11:42:53.508635 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.514114 kubelet[2999]: E0124 11:42:53.512513 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.514308 kubelet[2999]: W0124 11:42:53.514110 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.514308 kubelet[2999]: E0124 11:42:53.514147 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.516020 kubelet[2999]: E0124 11:42:53.515469 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.516020 kubelet[2999]: W0124 11:42:53.515555 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.516020 kubelet[2999]: E0124 11:42:53.515586 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.517378 kubelet[2999]: E0124 11:42:53.517142 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.517653 kubelet[2999]: W0124 11:42:53.517500 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.518315 kubelet[2999]: E0124 11:42:53.518283 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.519112 containerd[1625]: time="2026-01-24T11:42:53.519067742Z" level=info msg="connecting to shim ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622" address="unix:///run/containerd/s/d7081738d75c2d1d4c7baf15cfe6eef7456ccfa365fb6019cc7423ac316da361" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:42:53.520708 kubelet[2999]: E0124 11:42:53.520555 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.520791 kubelet[2999]: W0124 11:42:53.520626 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.520791 kubelet[2999]: E0124 11:42:53.520790 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.522919 kubelet[2999]: E0124 11:42:53.522641 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.522919 kubelet[2999]: W0124 11:42:53.522717 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.522919 kubelet[2999]: E0124 11:42:53.522735 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.525309 kubelet[2999]: E0124 11:42:53.525041 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.525309 kubelet[2999]: W0124 11:42:53.525173 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.525309 kubelet[2999]: E0124 11:42:53.525287 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.533533 kubelet[2999]: E0124 11:42:53.533343 2999 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:42:53.533533 kubelet[2999]: W0124 11:42:53.533372 2999 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:42:53.534319 kubelet[2999]: E0124 11:42:53.534043 2999 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:42:53.561598 containerd[1625]: time="2026-01-24T11:42:53.561408673Z" level=info msg="connecting to shim 4bb709a2de7e681839df049fb5e5ab54af60882f610077fe54c035e367d468f6" address="unix:///run/containerd/s/0064a95456e6e66a3297d123f2bd0073b4cdaede74210b84d1f2cb23d3a3f045" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:42:53.591662 systemd[1]: Started cri-containerd-ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622.scope - libcontainer container ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622. Jan 24 11:42:53.637000 audit: BPF prog-id=149 op=LOAD Jan 24 11:42:53.638000 audit: BPF prog-id=150 op=LOAD Jan 24 11:42:53.638000 audit[3526]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0238 a2=98 a3=0 items=0 ppid=3498 pid=3526 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.638000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365643238653465646632313739303038636265343363343462346565 Jan 24 11:42:53.638000 audit: BPF prog-id=150 op=UNLOAD Jan 24 11:42:53.638000 audit[3526]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=3526 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.638000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365643238653465646632313739303038636265343363343462346565 Jan 24 11:42:53.646000 audit: BPF prog-id=151 op=LOAD Jan 24 11:42:53.646000 audit[3526]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=3498 pid=3526 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.646000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365643238653465646632313739303038636265343363343462346565 Jan 24 11:42:53.652000 audit: BPF prog-id=152 op=LOAD Jan 24 11:42:53.652000 audit[3526]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=3498 pid=3526 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.652000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365643238653465646632313739303038636265343363343462346565 Jan 24 11:42:53.658000 audit: BPF prog-id=152 op=UNLOAD Jan 24 11:42:53.658000 audit[3526]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=3526 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.658000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365643238653465646632313739303038636265343363343462346565 Jan 24 11:42:53.658000 audit: BPF prog-id=151 op=UNLOAD Jan 24 11:42:53.658000 audit[3526]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=3526 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.658000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365643238653465646632313739303038636265343363343462346565 Jan 24 11:42:53.658000 audit: BPF prog-id=153 op=LOAD Jan 24 11:42:53.658000 audit[3526]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=3498 pid=3526 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.658000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365643238653465646632313739303038636265343363343462346565 Jan 24 11:42:53.666636 systemd[1]: Started cri-containerd-4bb709a2de7e681839df049fb5e5ab54af60882f610077fe54c035e367d468f6.scope - libcontainer container 4bb709a2de7e681839df049fb5e5ab54af60882f610077fe54c035e367d468f6. Jan 24 11:42:53.703000 audit: BPF prog-id=154 op=LOAD Jan 24 11:42:53.706000 audit: BPF prog-id=155 op=LOAD Jan 24 11:42:53.706000 audit[3560]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00018c238 a2=98 a3=0 items=0 ppid=3541 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.706000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3462623730396132646537653638313833396466303439666235653561 Jan 24 11:42:53.707000 audit: BPF prog-id=155 op=UNLOAD Jan 24 11:42:53.707000 audit[3560]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3541 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.707000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3462623730396132646537653638313833396466303439666235653561 Jan 24 11:42:53.707000 audit: BPF prog-id=156 op=LOAD Jan 24 11:42:53.707000 audit[3560]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00018c488 a2=98 a3=0 items=0 ppid=3541 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.707000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3462623730396132646537653638313833396466303439666235653561 Jan 24 11:42:53.708000 audit: BPF prog-id=157 op=LOAD Jan 24 11:42:53.708000 audit[3560]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00018c218 a2=98 a3=0 items=0 ppid=3541 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.708000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3462623730396132646537653638313833396466303439666235653561 Jan 24 11:42:53.708000 audit: BPF prog-id=157 op=UNLOAD Jan 24 11:42:53.708000 audit[3560]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3541 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.708000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3462623730396132646537653638313833396466303439666235653561 Jan 24 11:42:53.709000 audit: BPF prog-id=156 op=UNLOAD Jan 24 11:42:53.709000 audit[3560]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3541 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.709000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3462623730396132646537653638313833396466303439666235653561 Jan 24 11:42:53.709000 audit: BPF prog-id=158 op=LOAD Jan 24 11:42:53.709000 audit[3560]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00018c6e8 a2=98 a3=0 items=0 ppid=3541 pid=3560 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.709000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3462623730396132646537653638313833396466303439666235653561 Jan 24 11:42:53.733000 audit[3591]: NETFILTER_CFG table=filter:115 family=2 entries=22 op=nft_register_rule pid=3591 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:53.733000 audit[3591]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffd7c5dffc0 a2=0 a3=7ffd7c5dffac items=0 ppid=3157 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.733000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:53.740000 audit[3591]: NETFILTER_CFG table=nat:116 family=2 entries=12 op=nft_register_rule pid=3591 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:42:53.740000 audit[3591]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd7c5dffc0 a2=0 a3=0 items=0 ppid=3157 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:53.740000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:42:53.782761 containerd[1625]: time="2026-01-24T11:42:53.782338969Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-knsxr,Uid:49b6127d-07b4-4854-8332-f6cab36ded5e,Namespace:calico-system,Attempt:0,} returns sandbox id \"ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622\"" Jan 24 11:42:53.791719 kubelet[2999]: E0124 11:42:53.791671 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:53.805558 containerd[1625]: time="2026-01-24T11:42:53.804763862Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Jan 24 11:42:53.833336 containerd[1625]: time="2026-01-24T11:42:53.833021089Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-8dcb8f48d-jl7pj,Uid:adf2c65c-497b-4e5d-b8ba-e6f639ae408e,Namespace:calico-system,Attempt:0,} returns sandbox id \"4bb709a2de7e681839df049fb5e5ab54af60882f610077fe54c035e367d468f6\"" Jan 24 11:42:53.836138 kubelet[2999]: E0124 11:42:53.835136 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:54.662612 kubelet[2999]: E0124 11:42:54.662546 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:42:54.926740 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount778734741.mount: Deactivated successfully. Jan 24 11:42:55.248611 containerd[1625]: time="2026-01-24T11:42:55.248409461Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:42:55.252025 containerd[1625]: time="2026-01-24T11:42:55.251979772Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=0" Jan 24 11:42:55.256015 containerd[1625]: time="2026-01-24T11:42:55.255619024Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:42:55.267575 containerd[1625]: time="2026-01-24T11:42:55.265333330Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:42:55.267575 containerd[1625]: time="2026-01-24T11:42:55.266606569Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 1.461701245s" Jan 24 11:42:55.267575 containerd[1625]: time="2026-01-24T11:42:55.266639852Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Jan 24 11:42:55.272951 containerd[1625]: time="2026-01-24T11:42:55.272512623Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Jan 24 11:42:55.286302 containerd[1625]: time="2026-01-24T11:42:55.285916198Z" level=info msg="CreateContainer within sandbox \"ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jan 24 11:42:55.334252 containerd[1625]: time="2026-01-24T11:42:55.334099928Z" level=info msg="Container 1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:42:55.356863 containerd[1625]: time="2026-01-24T11:42:55.356608972Z" level=info msg="CreateContainer within sandbox \"ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151\"" Jan 24 11:42:55.361279 containerd[1625]: time="2026-01-24T11:42:55.359427526Z" level=info msg="StartContainer for \"1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151\"" Jan 24 11:42:55.364591 containerd[1625]: time="2026-01-24T11:42:55.364121084Z" level=info msg="connecting to shim 1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151" address="unix:///run/containerd/s/d7081738d75c2d1d4c7baf15cfe6eef7456ccfa365fb6019cc7423ac316da361" protocol=ttrpc version=3 Jan 24 11:42:55.485777 systemd[1]: Started cri-containerd-1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151.scope - libcontainer container 1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151. Jan 24 11:42:55.661000 audit: BPF prog-id=159 op=LOAD Jan 24 11:42:55.667522 kernel: kauditd_printk_skb: 58 callbacks suppressed Jan 24 11:42:55.667721 kernel: audit: type=1334 audit(1769254975.661:559): prog-id=159 op=LOAD Jan 24 11:42:55.661000 audit[3606]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3498 pid=3606 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:55.714113 kernel: audit: type=1300 audit(1769254975.661:559): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3498 pid=3606 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:55.714401 kernel: audit: type=1327 audit(1769254975.661:559): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3164313362383733366131636562323065616666346535613262363337 Jan 24 11:42:55.661000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3164313362383733366131636562323065616666346535613262363337 Jan 24 11:42:55.661000 audit: BPF prog-id=160 op=LOAD Jan 24 11:42:55.747269 kernel: audit: type=1334 audit(1769254975.661:560): prog-id=160 op=LOAD Jan 24 11:42:55.661000 audit[3606]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3498 pid=3606 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:55.776562 kernel: audit: type=1300 audit(1769254975.661:560): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3498 pid=3606 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:55.661000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3164313362383733366131636562323065616666346535613262363337 Jan 24 11:42:55.809285 kernel: audit: type=1327 audit(1769254975.661:560): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3164313362383733366131636562323065616666346535613262363337 Jan 24 11:42:55.661000 audit: BPF prog-id=160 op=UNLOAD Jan 24 11:42:55.661000 audit[3606]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=3606 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:55.834020 containerd[1625]: time="2026-01-24T11:42:55.833942931Z" level=info msg="StartContainer for \"1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151\" returns successfully" Jan 24 11:42:55.838741 kernel: audit: type=1334 audit(1769254975.661:561): prog-id=160 op=UNLOAD Jan 24 11:42:55.839961 kernel: audit: type=1300 audit(1769254975.661:561): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=3606 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:55.661000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3164313362383733366131636562323065616666346535613262363337 Jan 24 11:42:55.868040 kernel: audit: type=1327 audit(1769254975.661:561): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3164313362383733366131636562323065616666346535613262363337 Jan 24 11:42:55.869769 kernel: audit: type=1334 audit(1769254975.661:562): prog-id=159 op=UNLOAD Jan 24 11:42:55.661000 audit: BPF prog-id=159 op=UNLOAD Jan 24 11:42:55.661000 audit[3606]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=3606 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:55.661000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3164313362383733366131636562323065616666346535613262363337 Jan 24 11:42:55.661000 audit: BPF prog-id=161 op=LOAD Jan 24 11:42:55.661000 audit[3606]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=3498 pid=3606 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:42:55.661000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3164313362383733366131636562323065616666346535613262363337 Jan 24 11:42:55.887497 systemd[1]: cri-containerd-1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151.scope: Deactivated successfully. Jan 24 11:42:55.896000 audit: BPF prog-id=161 op=UNLOAD Jan 24 11:42:55.898620 containerd[1625]: time="2026-01-24T11:42:55.897578369Z" level=info msg="received container exit event container_id:\"1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151\" id:\"1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151\" pid:3622 exited_at:{seconds:1769254975 nanos:895542336}" Jan 24 11:42:56.028594 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1d13b8736a1ceb20eaff4e5a2b637008ac4dff5a4d9bfb2d5f7a561921f70151-rootfs.mount: Deactivated successfully. Jan 24 11:42:56.341375 kubelet[2999]: E0124 11:42:56.340987 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:42:56.662775 kubelet[2999]: E0124 11:42:56.662384 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:42:58.661995 kubelet[2999]: E0124 11:42:58.661681 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:00.554621 containerd[1625]: time="2026-01-24T11:43:00.554301316Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:43:00.560111 containerd[1625]: time="2026-01-24T11:43:00.559881574Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=33735893" Jan 24 11:43:00.564294 containerd[1625]: time="2026-01-24T11:43:00.563351535Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:43:00.573884 containerd[1625]: time="2026-01-24T11:43:00.573695177Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:43:00.575969 containerd[1625]: time="2026-01-24T11:43:00.575011856Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 5.302452235s" Jan 24 11:43:00.575969 containerd[1625]: time="2026-01-24T11:43:00.575110439Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Jan 24 11:43:00.581565 containerd[1625]: time="2026-01-24T11:43:00.581451413Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Jan 24 11:43:00.622790 containerd[1625]: time="2026-01-24T11:43:00.622719507Z" level=info msg="CreateContainer within sandbox \"4bb709a2de7e681839df049fb5e5ab54af60882f610077fe54c035e367d468f6\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jan 24 11:43:00.648396 containerd[1625]: time="2026-01-24T11:43:00.646418616Z" level=info msg="Container afa9be6d4f01f9db17a0673e623f4791eda1ad2950481cef43060092fdc038e6: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:43:00.663944 kubelet[2999]: E0124 11:43:00.663478 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:00.677758 containerd[1625]: time="2026-01-24T11:43:00.677571068Z" level=info msg="CreateContainer within sandbox \"4bb709a2de7e681839df049fb5e5ab54af60882f610077fe54c035e367d468f6\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"afa9be6d4f01f9db17a0673e623f4791eda1ad2950481cef43060092fdc038e6\"" Jan 24 11:43:00.679971 containerd[1625]: time="2026-01-24T11:43:00.679703643Z" level=info msg="StartContainer for \"afa9be6d4f01f9db17a0673e623f4791eda1ad2950481cef43060092fdc038e6\"" Jan 24 11:43:00.684198 containerd[1625]: time="2026-01-24T11:43:00.684132539Z" level=info msg="connecting to shim afa9be6d4f01f9db17a0673e623f4791eda1ad2950481cef43060092fdc038e6" address="unix:///run/containerd/s/0064a95456e6e66a3297d123f2bd0073b4cdaede74210b84d1f2cb23d3a3f045" protocol=ttrpc version=3 Jan 24 11:43:00.812782 systemd[1]: Started cri-containerd-afa9be6d4f01f9db17a0673e623f4791eda1ad2950481cef43060092fdc038e6.scope - libcontainer container afa9be6d4f01f9db17a0673e623f4791eda1ad2950481cef43060092fdc038e6. Jan 24 11:43:00.912000 audit: BPF prog-id=162 op=LOAD Jan 24 11:43:00.930501 kernel: kauditd_printk_skb: 6 callbacks suppressed Jan 24 11:43:00.930717 kernel: audit: type=1334 audit(1769254980.912:565): prog-id=162 op=LOAD Jan 24 11:43:00.914000 audit: BPF prog-id=163 op=LOAD Jan 24 11:43:00.914000 audit[3665]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:00.970928 kernel: audit: type=1334 audit(1769254980.914:566): prog-id=163 op=LOAD Jan 24 11:43:00.971059 kernel: audit: type=1300 audit(1769254980.914:566): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:00.914000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:01.019391 kernel: audit: type=1327 audit(1769254980.914:566): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:01.026629 kernel: audit: type=1334 audit(1769254980.914:567): prog-id=163 op=UNLOAD Jan 24 11:43:00.914000 audit: BPF prog-id=163 op=UNLOAD Jan 24 11:43:01.028752 kernel: audit: type=1300 audit(1769254980.914:567): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:00.914000 audit[3665]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:01.058410 kernel: audit: type=1327 audit(1769254980.914:567): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:00.914000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:00.914000 audit: BPF prog-id=164 op=LOAD Jan 24 11:43:01.088433 kernel: audit: type=1334 audit(1769254980.914:568): prog-id=164 op=LOAD Jan 24 11:43:00.914000 audit[3665]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:01.120156 containerd[1625]: time="2026-01-24T11:43:01.120016672Z" level=info msg="StartContainer for \"afa9be6d4f01f9db17a0673e623f4791eda1ad2950481cef43060092fdc038e6\" returns successfully" Jan 24 11:43:01.131771 kernel: audit: type=1300 audit(1769254980.914:568): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:00.914000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:00.914000 audit: BPF prog-id=165 op=LOAD Jan 24 11:43:00.914000 audit[3665]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:00.914000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:01.160323 kernel: audit: type=1327 audit(1769254980.914:568): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:00.915000 audit: BPF prog-id=165 op=UNLOAD Jan 24 11:43:00.915000 audit[3665]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:00.915000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:00.915000 audit: BPF prog-id=164 op=UNLOAD Jan 24 11:43:00.915000 audit[3665]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:00.915000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:00.915000 audit: BPF prog-id=166 op=LOAD Jan 24 11:43:00.915000 audit[3665]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=3541 pid=3665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:00.915000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166613962653664346630316639646231376130363733653632336634 Jan 24 11:43:01.437618 kubelet[2999]: E0124 11:43:01.432901 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:02.435726 kubelet[2999]: I0124 11:43:02.435437 2999 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 24 11:43:02.437168 kubelet[2999]: E0124 11:43:02.436146 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:02.664887 kubelet[2999]: E0124 11:43:02.663619 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:04.666096 kubelet[2999]: E0124 11:43:04.666017 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:06.667119 kubelet[2999]: E0124 11:43:06.666709 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:08.965447 kubelet[2999]: E0124 11:43:08.956758 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:09.952466 containerd[1625]: time="2026-01-24T11:43:09.952062478Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:43:09.957534 containerd[1625]: time="2026-01-24T11:43:09.957475608Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70443237" Jan 24 11:43:09.965330 containerd[1625]: time="2026-01-24T11:43:09.963671765Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:43:09.971048 containerd[1625]: time="2026-01-24T11:43:09.970989674Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:43:09.972443 containerd[1625]: time="2026-01-24T11:43:09.972154852Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 9.390662502s" Jan 24 11:43:09.972443 containerd[1625]: time="2026-01-24T11:43:09.972430255Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Jan 24 11:43:09.994088 containerd[1625]: time="2026-01-24T11:43:09.994036348Z" level=info msg="CreateContainer within sandbox \"ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 24 11:43:10.177309 containerd[1625]: time="2026-01-24T11:43:10.170326052Z" level=info msg="Container 16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:43:10.249486 containerd[1625]: time="2026-01-24T11:43:10.248616800Z" level=info msg="CreateContainer within sandbox \"ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413\"" Jan 24 11:43:10.253161 containerd[1625]: time="2026-01-24T11:43:10.253032626Z" level=info msg="StartContainer for \"16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413\"" Jan 24 11:43:10.258072 containerd[1625]: time="2026-01-24T11:43:10.257982606Z" level=info msg="connecting to shim 16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413" address="unix:///run/containerd/s/d7081738d75c2d1d4c7baf15cfe6eef7456ccfa365fb6019cc7423ac316da361" protocol=ttrpc version=3 Jan 24 11:43:10.559528 systemd[1]: Started cri-containerd-16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413.scope - libcontainer container 16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413. Jan 24 11:43:10.661782 kubelet[2999]: E0124 11:43:10.661701 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:10.725000 audit: BPF prog-id=167 op=LOAD Jan 24 11:43:10.732929 kernel: kauditd_printk_skb: 12 callbacks suppressed Jan 24 11:43:10.733045 kernel: audit: type=1334 audit(1769254990.725:573): prog-id=167 op=LOAD Jan 24 11:43:10.725000 audit[3714]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=3498 pid=3714 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:10.767331 kernel: audit: type=1300 audit(1769254990.725:573): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=3498 pid=3714 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:10.827777 kernel: audit: type=1327 audit(1769254990.725:573): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3136626661373832633964313535353830363834363935616534353234 Jan 24 11:43:10.828411 kernel: audit: type=1334 audit(1769254990.725:574): prog-id=168 op=LOAD Jan 24 11:43:10.828707 kernel: audit: type=1300 audit(1769254990.725:574): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=3498 pid=3714 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:10.725000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3136626661373832633964313535353830363834363935616534353234 Jan 24 11:43:10.725000 audit: BPF prog-id=168 op=LOAD Jan 24 11:43:10.725000 audit[3714]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=3498 pid=3714 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:10.725000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3136626661373832633964313535353830363834363935616534353234 Jan 24 11:43:10.726000 audit: BPF prog-id=168 op=UNLOAD Jan 24 11:43:10.877980 kernel: audit: type=1327 audit(1769254990.725:574): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3136626661373832633964313535353830363834363935616534353234 Jan 24 11:43:10.878078 kernel: audit: type=1334 audit(1769254990.726:575): prog-id=168 op=UNLOAD Jan 24 11:43:10.878138 kernel: audit: type=1300 audit(1769254990.726:575): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=3714 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:10.726000 audit[3714]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=3714 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:10.902682 kernel: audit: type=1327 audit(1769254990.726:575): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3136626661373832633964313535353830363834363935616534353234 Jan 24 11:43:10.726000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3136626661373832633964313535353830363834363935616534353234 Jan 24 11:43:10.726000 audit: BPF prog-id=167 op=UNLOAD Jan 24 11:43:10.931632 kernel: audit: type=1334 audit(1769254990.726:576): prog-id=167 op=UNLOAD Jan 24 11:43:10.726000 audit[3714]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=3714 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:10.726000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3136626661373832633964313535353830363834363935616534353234 Jan 24 11:43:10.726000 audit: BPF prog-id=169 op=LOAD Jan 24 11:43:10.726000 audit[3714]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=3498 pid=3714 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:10.726000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3136626661373832633964313535353830363834363935616534353234 Jan 24 11:43:10.983748 containerd[1625]: time="2026-01-24T11:43:10.983638462Z" level=info msg="StartContainer for \"16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413\" returns successfully" Jan 24 11:43:11.297450 kubelet[2999]: E0124 11:43:11.294025 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:11.892614 kubelet[2999]: E0124 11:43:11.891683 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:12.453089 kubelet[2999]: E0124 11:43:12.452518 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:12.480324 kubelet[2999]: I0124 11:43:12.478174 2999 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-8dcb8f48d-jl7pj" podStartSLOduration=13.724240426 podStartE2EDuration="20.466991526s" podCreationTimestamp="2026-01-24 11:42:52 +0000 UTC" firstStartedPulling="2026-01-24 11:42:53.838315868 +0000 UTC m=+29.637543055" lastFinishedPulling="2026-01-24 11:43:00.581066958 +0000 UTC m=+36.380294155" observedRunningTime="2026-01-24 11:43:01.535069783 +0000 UTC m=+37.334296971" watchObservedRunningTime="2026-01-24 11:43:12.466991526 +0000 UTC m=+48.266218713" Jan 24 11:43:13.666587 kubelet[2999]: E0124 11:43:13.665900 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:14.987668 systemd[1]: cri-containerd-16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413.scope: Deactivated successfully. Jan 24 11:43:14.991327 systemd[1]: cri-containerd-16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413.scope: Consumed 2.674s CPU time, 180.1M memory peak, 3.7M read from disk, 171.3M written to disk. Jan 24 11:43:14.993378 containerd[1625]: time="2026-01-24T11:43:14.993011149Z" level=info msg="received container exit event container_id:\"16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413\" id:\"16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413\" pid:3728 exited_at:{seconds:1769254994 nanos:992484993}" Jan 24 11:43:14.995000 audit: BPF prog-id=169 op=UNLOAD Jan 24 11:43:15.074659 kubelet[2999]: I0124 11:43:15.074554 2999 kubelet_node_status.go:439] "Fast updating node status as it just became ready" Jan 24 11:43:15.088351 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-16bfa782c9d155580684695ae4524de6d4a1615499c5893724bbcfc57b3aa413-rootfs.mount: Deactivated successfully. Jan 24 11:43:15.226602 systemd[1]: Created slice kubepods-burstable-pod6ef98b6a_4a5d_4999_ac24_f07f795beffa.slice - libcontainer container kubepods-burstable-pod6ef98b6a_4a5d_4999_ac24_f07f795beffa.slice. Jan 24 11:43:15.246332 systemd[1]: Created slice kubepods-burstable-podce8435a3_1751_4217_bed9_ea3dd2f1d3eb.slice - libcontainer container kubepods-burstable-podce8435a3_1751_4217_bed9_ea3dd2f1d3eb.slice. Jan 24 11:43:15.265375 systemd[1]: Created slice kubepods-besteffort-podec314a61_36d7_4af0_b9d6_11f6bfe86b8c.slice - libcontainer container kubepods-besteffort-podec314a61_36d7_4af0_b9d6_11f6bfe86b8c.slice. Jan 24 11:43:15.288478 systemd[1]: Created slice kubepods-besteffort-podeda31195_c93b_4ab6_93d0_16b77a14ef46.slice - libcontainer container kubepods-besteffort-podeda31195_c93b_4ab6_93d0_16b77a14ef46.slice. Jan 24 11:43:15.303975 systemd[1]: Created slice kubepods-besteffort-pod74372ec4_7697_478c_83d3_86a4742276e5.slice - libcontainer container kubepods-besteffort-pod74372ec4_7697_478c_83d3_86a4742276e5.slice. Jan 24 11:43:15.353318 kubelet[2999]: I0124 11:43:15.352614 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcbjk\" (UniqueName: \"kubernetes.io/projected/ce8435a3-1751-4217-bed9-ea3dd2f1d3eb-kube-api-access-zcbjk\") pod \"coredns-66bc5c9577-zx6pd\" (UID: \"ce8435a3-1751-4217-bed9-ea3dd2f1d3eb\") " pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:15.353318 kubelet[2999]: I0124 11:43:15.352742 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/eda31195-c93b-4ab6-93d0-16b77a14ef46-calico-apiserver-certs\") pod \"calico-apiserver-f58644cbc-dvnfm\" (UID: \"eda31195-c93b-4ab6-93d0-16b77a14ef46\") " pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:15.353318 kubelet[2999]: I0124 11:43:15.352778 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmcw2\" (UniqueName: \"kubernetes.io/projected/74372ec4-7697-478c-83d3-86a4742276e5-kube-api-access-xmcw2\") pod \"whisker-77c479b5bb-fngch\" (UID: \"74372ec4-7697-478c-83d3-86a4742276e5\") " pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:15.353318 kubelet[2999]: I0124 11:43:15.352924 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjcbb\" (UniqueName: \"kubernetes.io/projected/ec314a61-36d7-4af0-b9d6-11f6bfe86b8c-kube-api-access-bjcbb\") pod \"calico-apiserver-f58644cbc-hhctc\" (UID: \"ec314a61-36d7-4af0-b9d6-11f6bfe86b8c\") " pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" Jan 24 11:43:15.353318 kubelet[2999]: I0124 11:43:15.353046 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmffs\" (UniqueName: \"kubernetes.io/projected/eda31195-c93b-4ab6-93d0-16b77a14ef46-kube-api-access-pmffs\") pod \"calico-apiserver-f58644cbc-dvnfm\" (UID: \"eda31195-c93b-4ab6-93d0-16b77a14ef46\") " pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:15.354989 kubelet[2999]: I0124 11:43:15.353075 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/74372ec4-7697-478c-83d3-86a4742276e5-whisker-backend-key-pair\") pod \"whisker-77c479b5bb-fngch\" (UID: \"74372ec4-7697-478c-83d3-86a4742276e5\") " pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:15.354989 kubelet[2999]: I0124 11:43:15.353491 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/ec314a61-36d7-4af0-b9d6-11f6bfe86b8c-calico-apiserver-certs\") pod \"calico-apiserver-f58644cbc-hhctc\" (UID: \"ec314a61-36d7-4af0-b9d6-11f6bfe86b8c\") " pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" Jan 24 11:43:15.354989 kubelet[2999]: I0124 11:43:15.353529 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef98b6a-4a5d-4999-ac24-f07f795beffa-config-volume\") pod \"coredns-66bc5c9577-mlgbz\" (UID: \"6ef98b6a-4a5d-4999-ac24-f07f795beffa\") " pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:15.354989 kubelet[2999]: I0124 11:43:15.353551 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce8435a3-1751-4217-bed9-ea3dd2f1d3eb-config-volume\") pod \"coredns-66bc5c9577-zx6pd\" (UID: \"ce8435a3-1751-4217-bed9-ea3dd2f1d3eb\") " pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:15.354989 kubelet[2999]: I0124 11:43:15.353573 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74372ec4-7697-478c-83d3-86a4742276e5-whisker-ca-bundle\") pod \"whisker-77c479b5bb-fngch\" (UID: \"74372ec4-7697-478c-83d3-86a4742276e5\") " pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:15.355111 kubelet[2999]: I0124 11:43:15.353593 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l5dr\" (UniqueName: \"kubernetes.io/projected/6ef98b6a-4a5d-4999-ac24-f07f795beffa-kube-api-access-5l5dr\") pod \"coredns-66bc5c9577-mlgbz\" (UID: \"6ef98b6a-4a5d-4999-ac24-f07f795beffa\") " pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:15.372762 systemd[1]: Created slice kubepods-besteffort-pod9f91eb4d_a73b_4530_82b4_682e0b2b659a.slice - libcontainer container kubepods-besteffort-pod9f91eb4d_a73b_4530_82b4_682e0b2b659a.slice. Jan 24 11:43:15.378663 systemd[1]: Created slice kubepods-besteffort-podc101de33_34a1_459f_bcfc_2378435cf35e.slice - libcontainer container kubepods-besteffort-podc101de33_34a1_459f_bcfc_2378435cf35e.slice. Jan 24 11:43:15.455047 kubelet[2999]: I0124 11:43:15.454494 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c101de33-34a1-459f-bcfc-2378435cf35e-tigera-ca-bundle\") pod \"calico-kube-controllers-7dbccbb54b-mvbw5\" (UID: \"c101de33-34a1-459f-bcfc-2378435cf35e\") " pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" Jan 24 11:43:15.455047 kubelet[2999]: I0124 11:43:15.454618 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8dv5\" (UniqueName: \"kubernetes.io/projected/c101de33-34a1-459f-bcfc-2378435cf35e-kube-api-access-h8dv5\") pod \"calico-kube-controllers-7dbccbb54b-mvbw5\" (UID: \"c101de33-34a1-459f-bcfc-2378435cf35e\") " pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" Jan 24 11:43:15.455047 kubelet[2999]: I0124 11:43:15.454923 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f91eb4d-a73b-4530-82b4-682e0b2b659a-config\") pod \"goldmane-7c778bb748-4gb25\" (UID: \"9f91eb4d-a73b-4530-82b4-682e0b2b659a\") " pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:15.455047 kubelet[2999]: I0124 11:43:15.454950 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f91eb4d-a73b-4530-82b4-682e0b2b659a-goldmane-ca-bundle\") pod \"goldmane-7c778bb748-4gb25\" (UID: \"9f91eb4d-a73b-4530-82b4-682e0b2b659a\") " pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:15.455047 kubelet[2999]: I0124 11:43:15.454997 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w562n\" (UniqueName: \"kubernetes.io/projected/9f91eb4d-a73b-4530-82b4-682e0b2b659a-kube-api-access-w562n\") pod \"goldmane-7c778bb748-4gb25\" (UID: \"9f91eb4d-a73b-4530-82b4-682e0b2b659a\") " pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:15.460088 kubelet[2999]: I0124 11:43:15.455129 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/9f91eb4d-a73b-4530-82b4-682e0b2b659a-goldmane-key-pair\") pod \"goldmane-7c778bb748-4gb25\" (UID: \"9f91eb4d-a73b-4530-82b4-682e0b2b659a\") " pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:15.518475 kubelet[2999]: E0124 11:43:15.518153 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:15.522694 containerd[1625]: time="2026-01-24T11:43:15.522598523Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Jan 24 11:43:15.618430 containerd[1625]: time="2026-01-24T11:43:15.617962464Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:43:15.671319 containerd[1625]: time="2026-01-24T11:43:15.670483272Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77c479b5bb-fngch,Uid:74372ec4-7697-478c-83d3-86a4742276e5,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:15.694648 containerd[1625]: time="2026-01-24T11:43:15.694167066Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-4gb25,Uid:9f91eb4d-a73b-4530-82b4-682e0b2b659a,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:15.716666 systemd[1]: Created slice kubepods-besteffort-pod9501b5f5_3a29_45b9_8cd6_9e7a8c6911fd.slice - libcontainer container kubepods-besteffort-pod9501b5f5_3a29_45b9_8cd6_9e7a8c6911fd.slice. Jan 24 11:43:15.722358 containerd[1625]: time="2026-01-24T11:43:15.722138877Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7dbccbb54b-mvbw5,Uid:c101de33-34a1-459f-bcfc-2378435cf35e,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:15.746896 containerd[1625]: time="2026-01-24T11:43:15.746614013Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:15.846528 kubelet[2999]: E0124 11:43:15.845556 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:15.848901 containerd[1625]: time="2026-01-24T11:43:15.848552140Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,}" Jan 24 11:43:15.871115 kubelet[2999]: E0124 11:43:15.868959 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:15.885098 containerd[1625]: time="2026-01-24T11:43:15.884808850Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-hhctc,Uid:ec314a61-36d7-4af0-b9d6-11f6bfe86b8c,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:43:15.949992 containerd[1625]: time="2026-01-24T11:43:15.948770079Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,}" Jan 24 11:43:16.233159 containerd[1625]: time="2026-01-24T11:43:16.230896093Z" level=error msg="Failed to destroy network for sandbox \"d3b876fb88e921ef0eadc4307adc85a27f9bde04cdbe2b4bc2a703b9b190d978\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.236365 systemd[1]: run-netns-cni\x2d4aa99637\x2ddc5a\x2d41ec\x2dc079\x2dccbdb8e4d82c.mount: Deactivated successfully. Jan 24 11:43:16.268351 containerd[1625]: time="2026-01-24T11:43:16.268016166Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-4gb25,Uid:9f91eb4d-a73b-4530-82b4-682e0b2b659a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3b876fb88e921ef0eadc4307adc85a27f9bde04cdbe2b4bc2a703b9b190d978\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.270009 kubelet[2999]: E0124 11:43:16.269604 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3b876fb88e921ef0eadc4307adc85a27f9bde04cdbe2b4bc2a703b9b190d978\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.270009 kubelet[2999]: E0124 11:43:16.269728 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3b876fb88e921ef0eadc4307adc85a27f9bde04cdbe2b4bc2a703b9b190d978\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:16.270009 kubelet[2999]: E0124 11:43:16.269754 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d3b876fb88e921ef0eadc4307adc85a27f9bde04cdbe2b4bc2a703b9b190d978\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:16.273046 kubelet[2999]: E0124 11:43:16.269809 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d3b876fb88e921ef0eadc4307adc85a27f9bde04cdbe2b4bc2a703b9b190d978\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:43:16.319923 containerd[1625]: time="2026-01-24T11:43:16.317948451Z" level=error msg="Failed to destroy network for sandbox \"65102346a37960305cbc14a9fc94807baaf92ea3a8a3ae546b6261fab16f4b84\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.325330 systemd[1]: run-netns-cni\x2d1bb3e9f1\x2d0f47\x2db5e6\x2d4ce8\x2d17e2fd3659f3.mount: Deactivated successfully. Jan 24 11:43:16.375827 containerd[1625]: time="2026-01-24T11:43:16.375625734Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"65102346a37960305cbc14a9fc94807baaf92ea3a8a3ae546b6261fab16f4b84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.380087 kubelet[2999]: E0124 11:43:16.377055 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"65102346a37960305cbc14a9fc94807baaf92ea3a8a3ae546b6261fab16f4b84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.380087 kubelet[2999]: E0124 11:43:16.377141 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"65102346a37960305cbc14a9fc94807baaf92ea3a8a3ae546b6261fab16f4b84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:16.380087 kubelet[2999]: E0124 11:43:16.377166 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"65102346a37960305cbc14a9fc94807baaf92ea3a8a3ae546b6261fab16f4b84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:16.380407 kubelet[2999]: E0124 11:43:16.377328 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"65102346a37960305cbc14a9fc94807baaf92ea3a8a3ae546b6261fab16f4b84\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:43:16.412509 containerd[1625]: time="2026-01-24T11:43:16.411703888Z" level=error msg="Failed to destroy network for sandbox \"ca475576fa7a0eded8345e8515111ac12fe66897cf11d471c6c97a49acca0e60\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.422426 systemd[1]: run-netns-cni\x2d9f04515b\x2d1e09\x2d51b7\x2d2ab9\x2d110621445b46.mount: Deactivated successfully. Jan 24 11:43:16.438497 containerd[1625]: time="2026-01-24T11:43:16.438175003Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca475576fa7a0eded8345e8515111ac12fe66897cf11d471c6c97a49acca0e60\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.438816 kubelet[2999]: E0124 11:43:16.438779 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca475576fa7a0eded8345e8515111ac12fe66897cf11d471c6c97a49acca0e60\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.438948 kubelet[2999]: E0124 11:43:16.438898 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca475576fa7a0eded8345e8515111ac12fe66897cf11d471c6c97a49acca0e60\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:43:16.438948 kubelet[2999]: E0124 11:43:16.438922 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ca475576fa7a0eded8345e8515111ac12fe66897cf11d471c6c97a49acca0e60\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:43:16.438998 kubelet[2999]: E0124 11:43:16.438968 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ca475576fa7a0eded8345e8515111ac12fe66897cf11d471c6c97a49acca0e60\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:16.451057 containerd[1625]: time="2026-01-24T11:43:16.450906767Z" level=error msg="Failed to destroy network for sandbox \"373646d3e0d3efb800b2e4363626fe2ccb256d8537444325c02cfc9f8218791e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.464540 systemd[1]: run-netns-cni\x2d8fb3267f\x2d4d6d\x2d9b70\x2d3183\x2d2950132d5311.mount: Deactivated successfully. Jan 24 11:43:16.470503 containerd[1625]: time="2026-01-24T11:43:16.466813085Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7dbccbb54b-mvbw5,Uid:c101de33-34a1-459f-bcfc-2378435cf35e,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"373646d3e0d3efb800b2e4363626fe2ccb256d8537444325c02cfc9f8218791e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.470724 kubelet[2999]: E0124 11:43:16.467123 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"373646d3e0d3efb800b2e4363626fe2ccb256d8537444325c02cfc9f8218791e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.470724 kubelet[2999]: E0124 11:43:16.467301 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"373646d3e0d3efb800b2e4363626fe2ccb256d8537444325c02cfc9f8218791e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" Jan 24 11:43:16.470724 kubelet[2999]: E0124 11:43:16.467328 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"373646d3e0d3efb800b2e4363626fe2ccb256d8537444325c02cfc9f8218791e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" Jan 24 11:43:16.470973 kubelet[2999]: E0124 11:43:16.467380 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"373646d3e0d3efb800b2e4363626fe2ccb256d8537444325c02cfc9f8218791e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:43:16.478297 containerd[1625]: time="2026-01-24T11:43:16.477156101Z" level=error msg="Failed to destroy network for sandbox \"2a53e877959a742c43db86f9df7a8b87bd5ae81c3f23e1386fd2195440c8b011\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.493436 containerd[1625]: time="2026-01-24T11:43:16.493262632Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77c479b5bb-fngch,Uid:74372ec4-7697-478c-83d3-86a4742276e5,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a53e877959a742c43db86f9df7a8b87bd5ae81c3f23e1386fd2195440c8b011\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.497491 kubelet[2999]: E0124 11:43:16.496953 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a53e877959a742c43db86f9df7a8b87bd5ae81c3f23e1386fd2195440c8b011\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.497491 kubelet[2999]: E0124 11:43:16.497085 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a53e877959a742c43db86f9df7a8b87bd5ae81c3f23e1386fd2195440c8b011\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:16.497491 kubelet[2999]: E0124 11:43:16.497114 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a53e877959a742c43db86f9df7a8b87bd5ae81c3f23e1386fd2195440c8b011\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:16.497735 kubelet[2999]: E0124 11:43:16.497278 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-77c479b5bb-fngch_calico-system(74372ec4-7697-478c-83d3-86a4742276e5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-77c479b5bb-fngch_calico-system(74372ec4-7697-478c-83d3-86a4742276e5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2a53e877959a742c43db86f9df7a8b87bd5ae81c3f23e1386fd2195440c8b011\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-77c479b5bb-fngch" podUID="74372ec4-7697-478c-83d3-86a4742276e5" Jan 24 11:43:16.502645 containerd[1625]: time="2026-01-24T11:43:16.502597275Z" level=error msg="Failed to destroy network for sandbox \"55c96916147f6c42845cc14449d9a14d1827b68f2ac50ce5aa69e44d306b824b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.510967 containerd[1625]: time="2026-01-24T11:43:16.510404538Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"55c96916147f6c42845cc14449d9a14d1827b68f2ac50ce5aa69e44d306b824b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.511151 kubelet[2999]: E0124 11:43:16.510739 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"55c96916147f6c42845cc14449d9a14d1827b68f2ac50ce5aa69e44d306b824b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.511151 kubelet[2999]: E0124 11:43:16.510784 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"55c96916147f6c42845cc14449d9a14d1827b68f2ac50ce5aa69e44d306b824b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:16.511151 kubelet[2999]: E0124 11:43:16.510814 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"55c96916147f6c42845cc14449d9a14d1827b68f2ac50ce5aa69e44d306b824b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:16.511435 kubelet[2999]: E0124 11:43:16.511009 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-zx6pd_kube-system(ce8435a3-1751-4217-bed9-ea3dd2f1d3eb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-zx6pd_kube-system(ce8435a3-1751-4217-bed9-ea3dd2f1d3eb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"55c96916147f6c42845cc14449d9a14d1827b68f2ac50ce5aa69e44d306b824b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-zx6pd" podUID="ce8435a3-1751-4217-bed9-ea3dd2f1d3eb" Jan 24 11:43:16.520581 containerd[1625]: time="2026-01-24T11:43:16.520383999Z" level=error msg="Failed to destroy network for sandbox \"7b7230de511e709ebde9bb02db26203136f546463bd91c3506b21611df3db36a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.529632 containerd[1625]: time="2026-01-24T11:43:16.529301749Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b7230de511e709ebde9bb02db26203136f546463bd91c3506b21611df3db36a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.531322 kubelet[2999]: E0124 11:43:16.530512 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b7230de511e709ebde9bb02db26203136f546463bd91c3506b21611df3db36a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.531322 kubelet[2999]: E0124 11:43:16.531159 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b7230de511e709ebde9bb02db26203136f546463bd91c3506b21611df3db36a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:16.531431 kubelet[2999]: E0124 11:43:16.531342 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b7230de511e709ebde9bb02db26203136f546463bd91c3506b21611df3db36a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:16.532385 kubelet[2999]: E0124 11:43:16.531720 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-mlgbz_kube-system(6ef98b6a-4a5d-4999-ac24-f07f795beffa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-mlgbz_kube-system(6ef98b6a-4a5d-4999-ac24-f07f795beffa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7b7230de511e709ebde9bb02db26203136f546463bd91c3506b21611df3db36a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-mlgbz" podUID="6ef98b6a-4a5d-4999-ac24-f07f795beffa" Jan 24 11:43:16.580009 containerd[1625]: time="2026-01-24T11:43:16.579932436Z" level=error msg="Failed to destroy network for sandbox \"7316036764a13fe87b3aa9147ba7d13bfe0aa90eeb0a1437356b7b2d3492ef0a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.589623 containerd[1625]: time="2026-01-24T11:43:16.589547576Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-hhctc,Uid:ec314a61-36d7-4af0-b9d6-11f6bfe86b8c,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7316036764a13fe87b3aa9147ba7d13bfe0aa90eeb0a1437356b7b2d3492ef0a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.591390 kubelet[2999]: E0124 11:43:16.590962 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7316036764a13fe87b3aa9147ba7d13bfe0aa90eeb0a1437356b7b2d3492ef0a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:16.591390 kubelet[2999]: E0124 11:43:16.591110 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7316036764a13fe87b3aa9147ba7d13bfe0aa90eeb0a1437356b7b2d3492ef0a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" Jan 24 11:43:16.591390 kubelet[2999]: E0124 11:43:16.591152 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7316036764a13fe87b3aa9147ba7d13bfe0aa90eeb0a1437356b7b2d3492ef0a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" Jan 24 11:43:16.591655 kubelet[2999]: E0124 11:43:16.591360 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7316036764a13fe87b3aa9147ba7d13bfe0aa90eeb0a1437356b7b2d3492ef0a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:43:17.086968 systemd[1]: run-netns-cni\x2de1e0afe2\x2ddd02\x2de020\x2d572d\x2d228111d4974a.mount: Deactivated successfully. Jan 24 11:43:17.087455 systemd[1]: run-netns-cni\x2d61da4fc2\x2dcd51\x2d587d\x2dbc5f\x2d7a734e18ce16.mount: Deactivated successfully. Jan 24 11:43:17.087590 systemd[1]: run-netns-cni\x2d17d09937\x2d0d7a\x2d5957\x2dd95f\x2d768712b12e72.mount: Deactivated successfully. Jan 24 11:43:17.087706 systemd[1]: run-netns-cni\x2d7c104675\x2d20bc\x2d81b6\x2d9250\x2daf0fc29ae51d.mount: Deactivated successfully. Jan 24 11:43:26.899024 kubelet[2999]: E0124 11:43:26.897564 2999 kubelet.go:2617] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="8.229s" Jan 24 11:43:28.525814 kubelet[2999]: E0124 11:43:28.525134 2999 kubelet.go:2617] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.448s" Jan 24 11:43:28.546403 kubelet[2999]: E0124 11:43:28.543735 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:28.549308 kubelet[2999]: E0124 11:43:28.547767 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:28.555501 containerd[1625]: time="2026-01-24T11:43:28.553803242Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,}" Jan 24 11:43:28.560729 containerd[1625]: time="2026-01-24T11:43:28.555544686Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77c479b5bb-fngch,Uid:74372ec4-7697-478c-83d3-86a4742276e5,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:28.562286 containerd[1625]: time="2026-01-24T11:43:28.562102549Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-hhctc,Uid:ec314a61-36d7-4af0-b9d6-11f6bfe86b8c,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:43:28.832000 audit[4063]: NETFILTER_CFG table=filter:117 family=2 entries=21 op=nft_register_rule pid=4063 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:43:28.854164 kernel: kauditd_printk_skb: 6 callbacks suppressed Jan 24 11:43:28.854826 kernel: audit: type=1325 audit(1769255008.832:579): table=filter:117 family=2 entries=21 op=nft_register_rule pid=4063 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:43:28.832000 audit[4063]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc327f4340 a2=0 a3=7ffc327f432c items=0 ppid=3157 pid=4063 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:28.897470 kernel: audit: type=1300 audit(1769255008.832:579): arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc327f4340 a2=0 a3=7ffc327f432c items=0 ppid=3157 pid=4063 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:28.897603 kernel: audit: type=1327 audit(1769255008.832:579): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:43:28.832000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:43:28.861000 audit[4063]: NETFILTER_CFG table=nat:118 family=2 entries=19 op=nft_register_chain pid=4063 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:43:28.975004 kernel: audit: type=1325 audit(1769255008.861:580): table=nat:118 family=2 entries=19 op=nft_register_chain pid=4063 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:43:29.113833 kernel: audit: type=1300 audit(1769255008.861:580): arch=c000003e syscall=46 success=yes exit=6276 a0=3 a1=7ffc327f4340 a2=0 a3=7ffc327f432c items=0 ppid=3157 pid=4063 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:29.376721 kernel: audit: type=1327 audit(1769255008.861:580): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:43:28.861000 audit[4063]: SYSCALL arch=c000003e syscall=46 success=yes exit=6276 a0=3 a1=7ffc327f4340 a2=0 a3=7ffc327f432c items=0 ppid=3157 pid=4063 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:28.861000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:43:29.669724 containerd[1625]: time="2026-01-24T11:43:29.667476095Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7dbccbb54b-mvbw5,Uid:c101de33-34a1-459f-bcfc-2378435cf35e,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:29.673580 containerd[1625]: time="2026-01-24T11:43:29.673443603Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:29.677526 containerd[1625]: time="2026-01-24T11:43:29.677402309Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:43:29.765793 containerd[1625]: time="2026-01-24T11:43:29.765440806Z" level=error msg="Failed to destroy network for sandbox \"21430294eeb4a9e03e7f675511032ccbd37828e57443a9007d014c76307d3487\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:29.769794 systemd[1]: run-netns-cni\x2d8e6026fb\x2d7142\x2dece5\x2d7acf\x2d60b5319e788b.mount: Deactivated successfully. Jan 24 11:43:29.797396 containerd[1625]: time="2026-01-24T11:43:29.797067246Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-hhctc,Uid:ec314a61-36d7-4af0-b9d6-11f6bfe86b8c,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"21430294eeb4a9e03e7f675511032ccbd37828e57443a9007d014c76307d3487\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:29.804769 kubelet[2999]: E0124 11:43:29.804397 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"21430294eeb4a9e03e7f675511032ccbd37828e57443a9007d014c76307d3487\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:29.805623 kubelet[2999]: E0124 11:43:29.805055 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"21430294eeb4a9e03e7f675511032ccbd37828e57443a9007d014c76307d3487\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" Jan 24 11:43:29.805623 kubelet[2999]: E0124 11:43:29.805095 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"21430294eeb4a9e03e7f675511032ccbd37828e57443a9007d014c76307d3487\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" Jan 24 11:43:29.807020 kubelet[2999]: E0124 11:43:29.806649 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"21430294eeb4a9e03e7f675511032ccbd37828e57443a9007d014c76307d3487\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:43:29.861625 containerd[1625]: time="2026-01-24T11:43:29.861558143Z" level=error msg="Failed to destroy network for sandbox \"12816aa78d9defe0eb81d7ba2434d1571833344ac747c6f634cc4916bc98de3f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:29.868500 systemd[1]: run-netns-cni\x2def39e579\x2dfe59\x2d5b0f\x2db090\x2df2e0bd3e3852.mount: Deactivated successfully. Jan 24 11:43:29.897142 containerd[1625]: time="2026-01-24T11:43:29.897072231Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77c479b5bb-fngch,Uid:74372ec4-7697-478c-83d3-86a4742276e5,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"12816aa78d9defe0eb81d7ba2434d1571833344ac747c6f634cc4916bc98de3f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:29.901429 kubelet[2999]: E0124 11:43:29.901007 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"12816aa78d9defe0eb81d7ba2434d1571833344ac747c6f634cc4916bc98de3f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:29.901543 kubelet[2999]: E0124 11:43:29.901422 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"12816aa78d9defe0eb81d7ba2434d1571833344ac747c6f634cc4916bc98de3f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:29.901543 kubelet[2999]: E0124 11:43:29.901460 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"12816aa78d9defe0eb81d7ba2434d1571833344ac747c6f634cc4916bc98de3f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:29.902045 kubelet[2999]: E0124 11:43:29.901527 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-77c479b5bb-fngch_calico-system(74372ec4-7697-478c-83d3-86a4742276e5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-77c479b5bb-fngch_calico-system(74372ec4-7697-478c-83d3-86a4742276e5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"12816aa78d9defe0eb81d7ba2434d1571833344ac747c6f634cc4916bc98de3f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-77c479b5bb-fngch" podUID="74372ec4-7697-478c-83d3-86a4742276e5" Jan 24 11:43:29.965605 containerd[1625]: time="2026-01-24T11:43:29.965546309Z" level=error msg="Failed to destroy network for sandbox \"8bb90fce56db6c6fa2bdaf109c92b97b5c3976d65cb33360d1ac1d86bc2ebceb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:29.981050 containerd[1625]: time="2026-01-24T11:43:29.981002326Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8bb90fce56db6c6fa2bdaf109c92b97b5c3976d65cb33360d1ac1d86bc2ebceb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:29.984793 kubelet[2999]: E0124 11:43:29.984453 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8bb90fce56db6c6fa2bdaf109c92b97b5c3976d65cb33360d1ac1d86bc2ebceb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:29.984793 kubelet[2999]: E0124 11:43:29.984634 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8bb90fce56db6c6fa2bdaf109c92b97b5c3976d65cb33360d1ac1d86bc2ebceb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:29.984793 kubelet[2999]: E0124 11:43:29.984675 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8bb90fce56db6c6fa2bdaf109c92b97b5c3976d65cb33360d1ac1d86bc2ebceb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:29.985441 kubelet[2999]: E0124 11:43:29.984750 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-zx6pd_kube-system(ce8435a3-1751-4217-bed9-ea3dd2f1d3eb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-zx6pd_kube-system(ce8435a3-1751-4217-bed9-ea3dd2f1d3eb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8bb90fce56db6c6fa2bdaf109c92b97b5c3976d65cb33360d1ac1d86bc2ebceb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-zx6pd" podUID="ce8435a3-1751-4217-bed9-ea3dd2f1d3eb" Jan 24 11:43:30.267462 containerd[1625]: time="2026-01-24T11:43:30.266138836Z" level=error msg="Failed to destroy network for sandbox \"b0158a36c62d49a11463943284cf9a1e85975c0a191ee190fb81930c7709a0d1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:30.278837 containerd[1625]: time="2026-01-24T11:43:30.278690593Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7dbccbb54b-mvbw5,Uid:c101de33-34a1-459f-bcfc-2378435cf35e,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0158a36c62d49a11463943284cf9a1e85975c0a191ee190fb81930c7709a0d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:30.279712 kubelet[2999]: E0124 11:43:30.279672 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0158a36c62d49a11463943284cf9a1e85975c0a191ee190fb81930c7709a0d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:30.279990 kubelet[2999]: E0124 11:43:30.279959 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0158a36c62d49a11463943284cf9a1e85975c0a191ee190fb81930c7709a0d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" Jan 24 11:43:30.280299 kubelet[2999]: E0124 11:43:30.280087 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0158a36c62d49a11463943284cf9a1e85975c0a191ee190fb81930c7709a0d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" Jan 24 11:43:30.280998 kubelet[2999]: E0124 11:43:30.280508 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b0158a36c62d49a11463943284cf9a1e85975c0a191ee190fb81930c7709a0d1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:43:30.282575 containerd[1625]: time="2026-01-24T11:43:30.282454943Z" level=error msg="Failed to destroy network for sandbox \"9e7c59d064e2c20bc27b31cb8f303e87022dcd07c9ef3efbe5928a00b9ffc1ee\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:30.299473 containerd[1625]: time="2026-01-24T11:43:30.297527658Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9e7c59d064e2c20bc27b31cb8f303e87022dcd07c9ef3efbe5928a00b9ffc1ee\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:30.301713 kubelet[2999]: E0124 11:43:30.301108 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9e7c59d064e2c20bc27b31cb8f303e87022dcd07c9ef3efbe5928a00b9ffc1ee\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:30.301713 kubelet[2999]: E0124 11:43:30.301168 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9e7c59d064e2c20bc27b31cb8f303e87022dcd07c9ef3efbe5928a00b9ffc1ee\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:43:30.301713 kubelet[2999]: E0124 11:43:30.301326 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9e7c59d064e2c20bc27b31cb8f303e87022dcd07c9ef3efbe5928a00b9ffc1ee\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:43:30.302001 kubelet[2999]: E0124 11:43:30.301402 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9e7c59d064e2c20bc27b31cb8f303e87022dcd07c9ef3efbe5928a00b9ffc1ee\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:30.321500 containerd[1625]: time="2026-01-24T11:43:30.321437515Z" level=error msg="Failed to destroy network for sandbox \"bec040cc203d611ca27fbc0301d2e3d5d0271c553cb880afed2a1cc3c5e630b6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:30.327146 containerd[1625]: time="2026-01-24T11:43:30.327036117Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bec040cc203d611ca27fbc0301d2e3d5d0271c553cb880afed2a1cc3c5e630b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:30.327967 kubelet[2999]: E0124 11:43:30.327715 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bec040cc203d611ca27fbc0301d2e3d5d0271c553cb880afed2a1cc3c5e630b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:30.327967 kubelet[2999]: E0124 11:43:30.327784 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bec040cc203d611ca27fbc0301d2e3d5d0271c553cb880afed2a1cc3c5e630b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:30.327967 kubelet[2999]: E0124 11:43:30.327812 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bec040cc203d611ca27fbc0301d2e3d5d0271c553cb880afed2a1cc3c5e630b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:30.328111 kubelet[2999]: E0124 11:43:30.327956 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bec040cc203d611ca27fbc0301d2e3d5d0271c553cb880afed2a1cc3c5e630b6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:43:30.690532 systemd[1]: run-netns-cni\x2d9da553aa\x2d1525\x2df78a\x2d6450\x2d780d3ee36d68.mount: Deactivated successfully. Jan 24 11:43:30.691548 systemd[1]: run-netns-cni\x2d4efe2a3b\x2d9db4\x2d7673\x2ddd3e\x2df04f29bf9c40.mount: Deactivated successfully. Jan 24 11:43:30.691742 systemd[1]: run-netns-cni\x2d6f2ce4f1\x2d2625\x2db7dd\x2d99c4\x2d27ef4bc52959.mount: Deactivated successfully. Jan 24 11:43:30.691932 systemd[1]: run-netns-cni\x2dd4ca72d5\x2d8be6\x2dc480\x2dfe0f\x2daeb2506569ad.mount: Deactivated successfully. Jan 24 11:43:31.669837 kubelet[2999]: E0124 11:43:31.669502 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:31.671648 containerd[1625]: time="2026-01-24T11:43:31.671174764Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,}" Jan 24 11:43:31.678826 containerd[1625]: time="2026-01-24T11:43:31.678521361Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-4gb25,Uid:9f91eb4d-a73b-4530-82b4-682e0b2b659a,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:31.877018 containerd[1625]: time="2026-01-24T11:43:31.876797880Z" level=error msg="Failed to destroy network for sandbox \"9445bf527607f03d3f67527b56c6b66b81812abc60c1859e144798ab00000ea8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:31.882133 systemd[1]: run-netns-cni\x2d6d86be4b\x2dd51f\x2dfc04\x2d2de5\x2d5bafa3dedbba.mount: Deactivated successfully. Jan 24 11:43:31.893606 containerd[1625]: time="2026-01-24T11:43:31.892153672Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-4gb25,Uid:9f91eb4d-a73b-4530-82b4-682e0b2b659a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9445bf527607f03d3f67527b56c6b66b81812abc60c1859e144798ab00000ea8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:31.893960 kubelet[2999]: E0124 11:43:31.892701 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9445bf527607f03d3f67527b56c6b66b81812abc60c1859e144798ab00000ea8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:31.893960 kubelet[2999]: E0124 11:43:31.892776 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9445bf527607f03d3f67527b56c6b66b81812abc60c1859e144798ab00000ea8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:31.893960 kubelet[2999]: E0124 11:43:31.892798 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9445bf527607f03d3f67527b56c6b66b81812abc60c1859e144798ab00000ea8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:31.894124 kubelet[2999]: E0124 11:43:31.892915 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9445bf527607f03d3f67527b56c6b66b81812abc60c1859e144798ab00000ea8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:43:31.899727 containerd[1625]: time="2026-01-24T11:43:31.897773691Z" level=error msg="Failed to destroy network for sandbox \"6c5699ca2a6ac6f9ebbc62b421793da6dc5fdb4d5d276299f64587add3bb79ec\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:31.903163 systemd[1]: run-netns-cni\x2d61638bd4\x2de6ad\x2d6861\x2d1346\x2d8c303f3a104c.mount: Deactivated successfully. Jan 24 11:43:31.913619 containerd[1625]: time="2026-01-24T11:43:31.913324763Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c5699ca2a6ac6f9ebbc62b421793da6dc5fdb4d5d276299f64587add3bb79ec\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:31.914429 kubelet[2999]: E0124 11:43:31.914016 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c5699ca2a6ac6f9ebbc62b421793da6dc5fdb4d5d276299f64587add3bb79ec\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:31.914429 kubelet[2999]: E0124 11:43:31.914408 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c5699ca2a6ac6f9ebbc62b421793da6dc5fdb4d5d276299f64587add3bb79ec\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:31.914629 kubelet[2999]: E0124 11:43:31.914443 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c5699ca2a6ac6f9ebbc62b421793da6dc5fdb4d5d276299f64587add3bb79ec\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:31.914839 kubelet[2999]: E0124 11:43:31.914670 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-mlgbz_kube-system(6ef98b6a-4a5d-4999-ac24-f07f795beffa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-mlgbz_kube-system(6ef98b6a-4a5d-4999-ac24-f07f795beffa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6c5699ca2a6ac6f9ebbc62b421793da6dc5fdb4d5d276299f64587add3bb79ec\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-mlgbz" podUID="6ef98b6a-4a5d-4999-ac24-f07f795beffa" Jan 24 11:43:40.674597 kubelet[2999]: E0124 11:43:40.667036 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:40.677750 kubelet[2999]: E0124 11:43:40.669817 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:40.680645 kubelet[2999]: E0124 11:43:40.680475 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:40.683787 containerd[1625]: time="2026-01-24T11:43:40.683716505Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,}" Jan 24 11:43:41.160166 containerd[1625]: time="2026-01-24T11:43:41.155505292Z" level=error msg="Failed to destroy network for sandbox \"22415d61146ff034cda2fdc1541a02a56709b9bba5655fc3571f9547a14d175a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:41.178118 systemd[1]: run-netns-cni\x2deca1e6b2\x2d34fb\x2d8a18\x2d09ee\x2dd7245eb89659.mount: Deactivated successfully. Jan 24 11:43:41.185522 containerd[1625]: time="2026-01-24T11:43:41.184703564Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"22415d61146ff034cda2fdc1541a02a56709b9bba5655fc3571f9547a14d175a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:41.187607 kubelet[2999]: E0124 11:43:41.187454 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"22415d61146ff034cda2fdc1541a02a56709b9bba5655fc3571f9547a14d175a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:41.188063 kubelet[2999]: E0124 11:43:41.187811 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"22415d61146ff034cda2fdc1541a02a56709b9bba5655fc3571f9547a14d175a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:41.188063 kubelet[2999]: E0124 11:43:41.187970 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"22415d61146ff034cda2fdc1541a02a56709b9bba5655fc3571f9547a14d175a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:41.188063 kubelet[2999]: E0124 11:43:41.188037 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-zx6pd_kube-system(ce8435a3-1751-4217-bed9-ea3dd2f1d3eb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-zx6pd_kube-system(ce8435a3-1751-4217-bed9-ea3dd2f1d3eb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"22415d61146ff034cda2fdc1541a02a56709b9bba5655fc3571f9547a14d175a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-zx6pd" podUID="ce8435a3-1751-4217-bed9-ea3dd2f1d3eb" Jan 24 11:43:41.728153 containerd[1625]: time="2026-01-24T11:43:41.727543932Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:43:42.144383 containerd[1625]: time="2026-01-24T11:43:42.140131266Z" level=error msg="Failed to destroy network for sandbox \"627fe11008f600e8381fd8bfc2e1ac1338cd1a79e307b926a67ded00f4894344\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:42.152781 systemd[1]: run-netns-cni\x2d29afa606\x2da7ac\x2d8e2d\x2dedb2\x2dc49e671e6466.mount: Deactivated successfully. Jan 24 11:43:42.155485 containerd[1625]: time="2026-01-24T11:43:42.154604613Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"627fe11008f600e8381fd8bfc2e1ac1338cd1a79e307b926a67ded00f4894344\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:42.157471 kubelet[2999]: E0124 11:43:42.157117 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"627fe11008f600e8381fd8bfc2e1ac1338cd1a79e307b926a67ded00f4894344\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:42.158033 kubelet[2999]: E0124 11:43:42.157605 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"627fe11008f600e8381fd8bfc2e1ac1338cd1a79e307b926a67ded00f4894344\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:42.158033 kubelet[2999]: E0124 11:43:42.157634 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"627fe11008f600e8381fd8bfc2e1ac1338cd1a79e307b926a67ded00f4894344\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:42.158033 kubelet[2999]: E0124 11:43:42.157750 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"627fe11008f600e8381fd8bfc2e1ac1338cd1a79e307b926a67ded00f4894344\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:43:42.745291 containerd[1625]: time="2026-01-24T11:43:42.744683224Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77c479b5bb-fngch,Uid:74372ec4-7697-478c-83d3-86a4742276e5,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:42.752712 containerd[1625]: time="2026-01-24T11:43:42.750425034Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,}" Jan 24 11:43:42.752845 kubelet[2999]: E0124 11:43:42.748787 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:43.137481 containerd[1625]: time="2026-01-24T11:43:43.133724468Z" level=error msg="Failed to destroy network for sandbox \"9934e2d2fd47e2efe7a0e5d341d2321644baa11ec177b2f10abe099e32b5d194\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:43.140836 systemd[1]: run-netns-cni\x2de986b571\x2d105c\x2d176e\x2d5979\x2d0074de71e460.mount: Deactivated successfully. Jan 24 11:43:43.146491 containerd[1625]: time="2026-01-24T11:43:43.146139702Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77c479b5bb-fngch,Uid:74372ec4-7697-478c-83d3-86a4742276e5,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9934e2d2fd47e2efe7a0e5d341d2321644baa11ec177b2f10abe099e32b5d194\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:43.149627 kubelet[2999]: E0124 11:43:43.149391 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9934e2d2fd47e2efe7a0e5d341d2321644baa11ec177b2f10abe099e32b5d194\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:43.149627 kubelet[2999]: E0124 11:43:43.149453 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9934e2d2fd47e2efe7a0e5d341d2321644baa11ec177b2f10abe099e32b5d194\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:43.149627 kubelet[2999]: E0124 11:43:43.149473 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9934e2d2fd47e2efe7a0e5d341d2321644baa11ec177b2f10abe099e32b5d194\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:43.150451 kubelet[2999]: E0124 11:43:43.149521 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-77c479b5bb-fngch_calico-system(74372ec4-7697-478c-83d3-86a4742276e5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-77c479b5bb-fngch_calico-system(74372ec4-7697-478c-83d3-86a4742276e5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9934e2d2fd47e2efe7a0e5d341d2321644baa11ec177b2f10abe099e32b5d194\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-77c479b5bb-fngch" podUID="74372ec4-7697-478c-83d3-86a4742276e5" Jan 24 11:43:43.199949 containerd[1625]: time="2026-01-24T11:43:43.198948993Z" level=error msg="Failed to destroy network for sandbox \"e7ce04bba0ba4dbc486d4faee7211fb9de70fd5e03c4ef7ecba7e5bb088c0e04\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:43.205794 systemd[1]: run-netns-cni\x2d5c0b2a16\x2dbe11\x2df2ff\x2d7874\x2d2e4cd609f0ee.mount: Deactivated successfully. Jan 24 11:43:43.215441 containerd[1625]: time="2026-01-24T11:43:43.214746197Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e7ce04bba0ba4dbc486d4faee7211fb9de70fd5e03c4ef7ecba7e5bb088c0e04\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:43.215716 kubelet[2999]: E0124 11:43:43.215595 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e7ce04bba0ba4dbc486d4faee7211fb9de70fd5e03c4ef7ecba7e5bb088c0e04\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:43.215716 kubelet[2999]: E0124 11:43:43.215671 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e7ce04bba0ba4dbc486d4faee7211fb9de70fd5e03c4ef7ecba7e5bb088c0e04\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:43.215716 kubelet[2999]: E0124 11:43:43.215709 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e7ce04bba0ba4dbc486d4faee7211fb9de70fd5e03c4ef7ecba7e5bb088c0e04\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:43.216637 kubelet[2999]: E0124 11:43:43.215781 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-mlgbz_kube-system(6ef98b6a-4a5d-4999-ac24-f07f795beffa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-mlgbz_kube-system(6ef98b6a-4a5d-4999-ac24-f07f795beffa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e7ce04bba0ba4dbc486d4faee7211fb9de70fd5e03c4ef7ecba7e5bb088c0e04\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-mlgbz" podUID="6ef98b6a-4a5d-4999-ac24-f07f795beffa" Jan 24 11:43:44.816108 containerd[1625]: time="2026-01-24T11:43:44.810861950Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7dbccbb54b-mvbw5,Uid:c101de33-34a1-459f-bcfc-2378435cf35e,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:44.816108 containerd[1625]: time="2026-01-24T11:43:44.860646559Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-4gb25,Uid:9f91eb4d-a73b-4530-82b4-682e0b2b659a,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:44.816108 containerd[1625]: time="2026-01-24T11:43:44.860764068Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:45.674422 kubelet[2999]: E0124 11:43:45.664438 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:45.703732 containerd[1625]: time="2026-01-24T11:43:45.702688708Z" level=error msg="Failed to destroy network for sandbox \"e6a47df5109a95d47114e6827b5c1b54dab0dcfba8fca9e7e3953435b5da1d3c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:45.722365 systemd[1]: run-netns-cni\x2d90296cd1\x2d500a\x2d3eef\x2d49e2\x2d0be520c6c3ca.mount: Deactivated successfully. Jan 24 11:43:45.835119 containerd[1625]: time="2026-01-24T11:43:45.831129778Z" level=error msg="Failed to destroy network for sandbox \"3f3c1fbeebb7610252c77f248347c7ae0395acd2d9f413a86a1514b3b6ea95b9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:45.839511 containerd[1625]: time="2026-01-24T11:43:45.839459865Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-hhctc,Uid:ec314a61-36d7-4af0-b9d6-11f6bfe86b8c,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:43:45.849054 systemd[1]: run-netns-cni\x2dd5dc04f8\x2d168f\x2d1fb0\x2d10f1\x2dd16d71e776d9.mount: Deactivated successfully. Jan 24 11:43:45.854981 containerd[1625]: time="2026-01-24T11:43:45.854627179Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7dbccbb54b-mvbw5,Uid:c101de33-34a1-459f-bcfc-2378435cf35e,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e6a47df5109a95d47114e6827b5c1b54dab0dcfba8fca9e7e3953435b5da1d3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:45.857705 kubelet[2999]: E0124 11:43:45.856674 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e6a47df5109a95d47114e6827b5c1b54dab0dcfba8fca9e7e3953435b5da1d3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:45.857705 kubelet[2999]: E0124 11:43:45.857100 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e6a47df5109a95d47114e6827b5c1b54dab0dcfba8fca9e7e3953435b5da1d3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" Jan 24 11:43:45.859167 kubelet[2999]: E0124 11:43:45.858821 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e6a47df5109a95d47114e6827b5c1b54dab0dcfba8fca9e7e3953435b5da1d3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" Jan 24 11:43:45.861473 kubelet[2999]: E0124 11:43:45.861375 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e6a47df5109a95d47114e6827b5c1b54dab0dcfba8fca9e7e3953435b5da1d3c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:43:45.931652 containerd[1625]: time="2026-01-24T11:43:45.928005260Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f3c1fbeebb7610252c77f248347c7ae0395acd2d9f413a86a1514b3b6ea95b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:45.939374 kubelet[2999]: E0124 11:43:45.936860 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f3c1fbeebb7610252c77f248347c7ae0395acd2d9f413a86a1514b3b6ea95b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:45.939374 kubelet[2999]: E0124 11:43:45.938166 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f3c1fbeebb7610252c77f248347c7ae0395acd2d9f413a86a1514b3b6ea95b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:43:45.939374 kubelet[2999]: E0124 11:43:45.939015 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f3c1fbeebb7610252c77f248347c7ae0395acd2d9f413a86a1514b3b6ea95b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:43:45.942826 kubelet[2999]: E0124 11:43:45.939167 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3f3c1fbeebb7610252c77f248347c7ae0395acd2d9f413a86a1514b3b6ea95b9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:46.000607 containerd[1625]: time="2026-01-24T11:43:46.000542412Z" level=error msg="Failed to destroy network for sandbox \"02475dbef256b5e751d6f08a82d4f0b8f86309680091d6d3366eb5b7a94514bf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:46.020708 systemd[1]: run-netns-cni\x2daa267a86\x2d4ea6\x2d7929\x2dbd43\x2df713f3669e5f.mount: Deactivated successfully. Jan 24 11:43:46.047165 containerd[1625]: time="2026-01-24T11:43:46.046310598Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-4gb25,Uid:9f91eb4d-a73b-4530-82b4-682e0b2b659a,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"02475dbef256b5e751d6f08a82d4f0b8f86309680091d6d3366eb5b7a94514bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:46.048448 kubelet[2999]: E0124 11:43:46.047324 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"02475dbef256b5e751d6f08a82d4f0b8f86309680091d6d3366eb5b7a94514bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:46.048448 kubelet[2999]: E0124 11:43:46.047405 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"02475dbef256b5e751d6f08a82d4f0b8f86309680091d6d3366eb5b7a94514bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:46.048448 kubelet[2999]: E0124 11:43:46.047432 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"02475dbef256b5e751d6f08a82d4f0b8f86309680091d6d3366eb5b7a94514bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-4gb25" Jan 24 11:43:46.048653 kubelet[2999]: E0124 11:43:46.047501 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"02475dbef256b5e751d6f08a82d4f0b8f86309680091d6d3366eb5b7a94514bf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:43:46.347134 containerd[1625]: time="2026-01-24T11:43:46.332621263Z" level=error msg="Failed to destroy network for sandbox \"609d01ab63c52fb085b667947b0e9e066dcf02379d47e5e2f8d8f957bf41d3d6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:46.364571 systemd[1]: run-netns-cni\x2d5633158c\x2d7efb\x2dc88a\x2dd148\x2d8aa6eff4c3fa.mount: Deactivated successfully. Jan 24 11:43:46.375402 containerd[1625]: time="2026-01-24T11:43:46.374496131Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-hhctc,Uid:ec314a61-36d7-4af0-b9d6-11f6bfe86b8c,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"609d01ab63c52fb085b667947b0e9e066dcf02379d47e5e2f8d8f957bf41d3d6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:46.382343 kubelet[2999]: E0124 11:43:46.381706 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"609d01ab63c52fb085b667947b0e9e066dcf02379d47e5e2f8d8f957bf41d3d6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:46.382658 kubelet[2999]: E0124 11:43:46.382623 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"609d01ab63c52fb085b667947b0e9e066dcf02379d47e5e2f8d8f957bf41d3d6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" Jan 24 11:43:46.382789 kubelet[2999]: E0124 11:43:46.382764 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"609d01ab63c52fb085b667947b0e9e066dcf02379d47e5e2f8d8f957bf41d3d6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" Jan 24 11:43:46.387595 kubelet[2999]: E0124 11:43:46.387456 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"609d01ab63c52fb085b667947b0e9e066dcf02379d47e5e2f8d8f957bf41d3d6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:43:47.664344 kubelet[2999]: E0124 11:43:47.660874 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:52.678974 containerd[1625]: time="2026-01-24T11:43:52.678706982Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:43:53.086469 containerd[1625]: time="2026-01-24T11:43:53.083044028Z" level=error msg="Failed to destroy network for sandbox \"d8ce0a857403995ecbd30b8ec48bcf6fe0ec2b1e71220cb236b928b4af6eaf31\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:53.087574 systemd[1]: run-netns-cni\x2d0cf4c795\x2d86aa\x2d07bc\x2dd7ef\x2d0c973aecdd5f.mount: Deactivated successfully. Jan 24 11:43:53.119053 containerd[1625]: time="2026-01-24T11:43:53.118980777Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d8ce0a857403995ecbd30b8ec48bcf6fe0ec2b1e71220cb236b928b4af6eaf31\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:53.119751 kubelet[2999]: E0124 11:43:53.119654 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d8ce0a857403995ecbd30b8ec48bcf6fe0ec2b1e71220cb236b928b4af6eaf31\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:53.120832 kubelet[2999]: E0124 11:43:53.119747 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d8ce0a857403995ecbd30b8ec48bcf6fe0ec2b1e71220cb236b928b4af6eaf31\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:53.120832 kubelet[2999]: E0124 11:43:53.119778 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d8ce0a857403995ecbd30b8ec48bcf6fe0ec2b1e71220cb236b928b4af6eaf31\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" Jan 24 11:43:53.120832 kubelet[2999]: E0124 11:43:53.119850 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d8ce0a857403995ecbd30b8ec48bcf6fe0ec2b1e71220cb236b928b4af6eaf31\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:43:53.676057 containerd[1625]: time="2026-01-24T11:43:53.675744569Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77c479b5bb-fngch,Uid:74372ec4-7697-478c-83d3-86a4742276e5,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:54.105533 containerd[1625]: time="2026-01-24T11:43:54.096486090Z" level=error msg="Failed to destroy network for sandbox \"bdb3e53bdeb3c73cd5bfc4c6fd1f05d423157b239179cd778c7e7056926f24a6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:54.124534 systemd[1]: run-netns-cni\x2d14750b38\x2da64c\x2d1475\x2da008\x2d6b754d1aed79.mount: Deactivated successfully. Jan 24 11:43:54.144556 containerd[1625]: time="2026-01-24T11:43:54.144434919Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-77c479b5bb-fngch,Uid:74372ec4-7697-478c-83d3-86a4742276e5,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bdb3e53bdeb3c73cd5bfc4c6fd1f05d423157b239179cd778c7e7056926f24a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:54.146778 kubelet[2999]: E0124 11:43:54.146551 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bdb3e53bdeb3c73cd5bfc4c6fd1f05d423157b239179cd778c7e7056926f24a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:54.146778 kubelet[2999]: E0124 11:43:54.146708 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bdb3e53bdeb3c73cd5bfc4c6fd1f05d423157b239179cd778c7e7056926f24a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:54.146778 kubelet[2999]: E0124 11:43:54.146740 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bdb3e53bdeb3c73cd5bfc4c6fd1f05d423157b239179cd778c7e7056926f24a6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-77c479b5bb-fngch" Jan 24 11:43:54.147692 kubelet[2999]: E0124 11:43:54.146805 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-77c479b5bb-fngch_calico-system(74372ec4-7697-478c-83d3-86a4742276e5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-77c479b5bb-fngch_calico-system(74372ec4-7697-478c-83d3-86a4742276e5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bdb3e53bdeb3c73cd5bfc4c6fd1f05d423157b239179cd778c7e7056926f24a6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-77c479b5bb-fngch" podUID="74372ec4-7697-478c-83d3-86a4742276e5" Jan 24 11:43:55.681363 kubelet[2999]: E0124 11:43:55.680481 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:55.688016 kubelet[2999]: E0124 11:43:55.687659 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:55.703623 containerd[1625]: time="2026-01-24T11:43:55.703568927Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,}" Jan 24 11:43:55.721636 containerd[1625]: time="2026-01-24T11:43:55.703779745Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,}" Jan 24 11:43:56.253477 containerd[1625]: time="2026-01-24T11:43:56.252869939Z" level=error msg="Failed to destroy network for sandbox \"d56e704bd4594704fddd7eefc8ca3f2147aaf1316a1e679e71c461d136e395fa\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:56.262500 systemd[1]: run-netns-cni\x2de49b2ee5\x2dd11c\x2d4faa\x2d70b2\x2da0415d13204c.mount: Deactivated successfully. Jan 24 11:43:56.269035 containerd[1625]: time="2026-01-24T11:43:56.268720832Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d56e704bd4594704fddd7eefc8ca3f2147aaf1316a1e679e71c461d136e395fa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:56.270642 kubelet[2999]: E0124 11:43:56.270595 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d56e704bd4594704fddd7eefc8ca3f2147aaf1316a1e679e71c461d136e395fa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:56.271777 kubelet[2999]: E0124 11:43:56.271743 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d56e704bd4594704fddd7eefc8ca3f2147aaf1316a1e679e71c461d136e395fa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:56.271984 kubelet[2999]: E0124 11:43:56.271869 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d56e704bd4594704fddd7eefc8ca3f2147aaf1316a1e679e71c461d136e395fa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-mlgbz" Jan 24 11:43:56.278317 kubelet[2999]: E0124 11:43:56.276484 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-mlgbz_kube-system(6ef98b6a-4a5d-4999-ac24-f07f795beffa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-mlgbz_kube-system(6ef98b6a-4a5d-4999-ac24-f07f795beffa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d56e704bd4594704fddd7eefc8ca3f2147aaf1316a1e679e71c461d136e395fa\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-mlgbz" podUID="6ef98b6a-4a5d-4999-ac24-f07f795beffa" Jan 24 11:43:56.340509 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount433301416.mount: Deactivated successfully. Jan 24 11:43:56.359157 containerd[1625]: time="2026-01-24T11:43:56.359092540Z" level=error msg="Failed to destroy network for sandbox \"94a1287476926f1ea6f4cabe9287bdb49f482e2463e40e953be6a4715782e4d5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:56.368837 systemd[1]: run-netns-cni\x2de041b85e\x2df829\x2d8f8b\x2d377d\x2dc84c07627d98.mount: Deactivated successfully. Jan 24 11:43:56.443470 containerd[1625]: time="2026-01-24T11:43:56.443314574Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"94a1287476926f1ea6f4cabe9287bdb49f482e2463e40e953be6a4715782e4d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:56.445523 kubelet[2999]: E0124 11:43:56.445371 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"94a1287476926f1ea6f4cabe9287bdb49f482e2463e40e953be6a4715782e4d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:56.445523 kubelet[2999]: E0124 11:43:56.445505 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"94a1287476926f1ea6f4cabe9287bdb49f482e2463e40e953be6a4715782e4d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:56.446040 kubelet[2999]: E0124 11:43:56.445535 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"94a1287476926f1ea6f4cabe9287bdb49f482e2463e40e953be6a4715782e4d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-zx6pd" Jan 24 11:43:56.446523 kubelet[2999]: E0124 11:43:56.446177 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-zx6pd_kube-system(ce8435a3-1751-4217-bed9-ea3dd2f1d3eb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-zx6pd_kube-system(ce8435a3-1751-4217-bed9-ea3dd2f1d3eb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"94a1287476926f1ea6f4cabe9287bdb49f482e2463e40e953be6a4715782e4d5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-zx6pd" podUID="ce8435a3-1751-4217-bed9-ea3dd2f1d3eb" Jan 24 11:43:56.451747 containerd[1625]: time="2026-01-24T11:43:56.451696693Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:43:56.455762 containerd[1625]: time="2026-01-24T11:43:56.455156532Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156880555" Jan 24 11:43:56.463771 containerd[1625]: time="2026-01-24T11:43:56.463105621Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:43:56.472995 containerd[1625]: time="2026-01-24T11:43:56.472957822Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:43:56.475611 containerd[1625]: time="2026-01-24T11:43:56.475437456Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 40.95273604s" Jan 24 11:43:56.475611 containerd[1625]: time="2026-01-24T11:43:56.475557640Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Jan 24 11:43:56.576504 containerd[1625]: time="2026-01-24T11:43:56.575129947Z" level=info msg="CreateContainer within sandbox \"ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jan 24 11:43:56.626730 containerd[1625]: time="2026-01-24T11:43:56.626674141Z" level=info msg="Container 5031554325f0b34d0ee5e283260d2a1f375ad41d298061458d16ae5499c9ee2b: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:43:56.687459 containerd[1625]: time="2026-01-24T11:43:56.686846004Z" level=info msg="CreateContainer within sandbox \"ced28e4edf2179008cbe43c44b4eed4ab4243e54111711297dac7abf477b9622\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"5031554325f0b34d0ee5e283260d2a1f375ad41d298061458d16ae5499c9ee2b\"" Jan 24 11:43:56.688875 containerd[1625]: time="2026-01-24T11:43:56.688753451Z" level=info msg="StartContainer for \"5031554325f0b34d0ee5e283260d2a1f375ad41d298061458d16ae5499c9ee2b\"" Jan 24 11:43:56.715828 containerd[1625]: time="2026-01-24T11:43:56.715401100Z" level=info msg="connecting to shim 5031554325f0b34d0ee5e283260d2a1f375ad41d298061458d16ae5499c9ee2b" address="unix:///run/containerd/s/d7081738d75c2d1d4c7baf15cfe6eef7456ccfa365fb6019cc7423ac316da361" protocol=ttrpc version=3 Jan 24 11:43:56.944631 systemd[1]: Started cri-containerd-5031554325f0b34d0ee5e283260d2a1f375ad41d298061458d16ae5499c9ee2b.scope - libcontainer container 5031554325f0b34d0ee5e283260d2a1f375ad41d298061458d16ae5499c9ee2b. Jan 24 11:43:57.126000 audit: BPF prog-id=170 op=LOAD Jan 24 11:43:57.137562 kernel: audit: type=1334 audit(1769255037.126:581): prog-id=170 op=LOAD Jan 24 11:43:57.126000 audit[4682]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3498 pid=4682 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:57.165483 kernel: audit: type=1300 audit(1769255037.126:581): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3498 pid=4682 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:57.126000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3530333135353433323566306233346430656535653238333236306432 Jan 24 11:43:57.225609 kernel: audit: type=1327 audit(1769255037.126:581): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3530333135353433323566306233346430656535653238333236306432 Jan 24 11:43:57.127000 audit: BPF prog-id=171 op=LOAD Jan 24 11:43:57.127000 audit[4682]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3498 pid=4682 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:57.262362 kernel: audit: type=1334 audit(1769255037.127:582): prog-id=171 op=LOAD Jan 24 11:43:57.262492 kernel: audit: type=1300 audit(1769255037.127:582): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3498 pid=4682 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:57.127000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3530333135353433323566306233346430656535653238333236306432 Jan 24 11:43:57.297435 kernel: audit: type=1327 audit(1769255037.127:582): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3530333135353433323566306233346430656535653238333236306432 Jan 24 11:43:57.297571 kernel: audit: type=1334 audit(1769255037.127:583): prog-id=171 op=UNLOAD Jan 24 11:43:57.127000 audit: BPF prog-id=171 op=UNLOAD Jan 24 11:43:57.127000 audit[4682]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=4682 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:57.347784 containerd[1625]: time="2026-01-24T11:43:57.347649792Z" level=info msg="StartContainer for \"5031554325f0b34d0ee5e283260d2a1f375ad41d298061458d16ae5499c9ee2b\" returns successfully" Jan 24 11:43:57.349450 kernel: audit: type=1300 audit(1769255037.127:583): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=4682 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:57.127000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3530333135353433323566306233346430656535653238333236306432 Jan 24 11:43:57.380442 kernel: audit: type=1327 audit(1769255037.127:583): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3530333135353433323566306233346430656535653238333236306432 Jan 24 11:43:57.127000 audit: BPF prog-id=170 op=UNLOAD Jan 24 11:43:57.389349 kernel: audit: type=1334 audit(1769255037.127:584): prog-id=170 op=UNLOAD Jan 24 11:43:57.127000 audit[4682]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3498 pid=4682 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:57.127000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3530333135353433323566306233346430656535653238333236306432 Jan 24 11:43:57.127000 audit: BPF prog-id=172 op=LOAD Jan 24 11:43:57.127000 audit[4682]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=3498 pid=4682 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:43:57.127000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3530333135353433323566306233346430656535653238333236306432 Jan 24 11:43:57.674695 containerd[1625]: time="2026-01-24T11:43:57.674642709Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:57.895509 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jan 24 11:43:57.895678 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jan 24 11:43:58.134093 containerd[1625]: time="2026-01-24T11:43:58.129404576Z" level=error msg="Failed to destroy network for sandbox \"4859e8f1178fa6ea90abe1210ed3767e1a000de4f8bb80abf59d1bbb11e3fa53\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:58.138325 systemd[1]: run-netns-cni\x2d20910e90\x2d81d9\x2dd041\x2d1e6d\x2d70d295853462.mount: Deactivated successfully. Jan 24 11:43:58.165781 containerd[1625]: time="2026-01-24T11:43:58.163869392Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4859e8f1178fa6ea90abe1210ed3767e1a000de4f8bb80abf59d1bbb11e3fa53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:58.166456 kubelet[2999]: E0124 11:43:58.164434 2999 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4859e8f1178fa6ea90abe1210ed3767e1a000de4f8bb80abf59d1bbb11e3fa53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:43:58.166456 kubelet[2999]: E0124 11:43:58.164519 2999 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4859e8f1178fa6ea90abe1210ed3767e1a000de4f8bb80abf59d1bbb11e3fa53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:43:58.166456 kubelet[2999]: E0124 11:43:58.164551 2999 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4859e8f1178fa6ea90abe1210ed3767e1a000de4f8bb80abf59d1bbb11e3fa53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j5qxv" Jan 24 11:43:58.167737 kubelet[2999]: E0124 11:43:58.164692 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4859e8f1178fa6ea90abe1210ed3767e1a000de4f8bb80abf59d1bbb11e3fa53\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:43:58.178854 kubelet[2999]: E0124 11:43:58.178638 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:58.688675 containerd[1625]: time="2026-01-24T11:43:58.688526569Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-hhctc,Uid:ec314a61-36d7-4af0-b9d6-11f6bfe86b8c,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:43:58.701056 containerd[1625]: time="2026-01-24T11:43:58.700167688Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-4gb25,Uid:9f91eb4d-a73b-4530-82b4-682e0b2b659a,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:58.722061 containerd[1625]: time="2026-01-24T11:43:58.721819788Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7dbccbb54b-mvbw5,Uid:c101de33-34a1-459f-bcfc-2378435cf35e,Namespace:calico-system,Attempt:0,}" Jan 24 11:43:58.980999 kubelet[2999]: I0124 11:43:58.977141 2999 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-knsxr" podStartSLOduration=4.281097882 podStartE2EDuration="1m6.97711472s" podCreationTimestamp="2026-01-24 11:42:52 +0000 UTC" firstStartedPulling="2026-01-24 11:42:53.803728521 +0000 UTC m=+29.602955709" lastFinishedPulling="2026-01-24 11:43:56.49974536 +0000 UTC m=+92.298972547" observedRunningTime="2026-01-24 11:43:58.389097871 +0000 UTC m=+94.188325089" watchObservedRunningTime="2026-01-24 11:43:58.97711472 +0000 UTC m=+94.776341907" Jan 24 11:43:59.095505 kubelet[2999]: I0124 11:43:59.093148 2999 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmcw2\" (UniqueName: \"kubernetes.io/projected/74372ec4-7697-478c-83d3-86a4742276e5-kube-api-access-xmcw2\") pod \"74372ec4-7697-478c-83d3-86a4742276e5\" (UID: \"74372ec4-7697-478c-83d3-86a4742276e5\") " Jan 24 11:43:59.095505 kubelet[2999]: I0124 11:43:59.093449 2999 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74372ec4-7697-478c-83d3-86a4742276e5-whisker-ca-bundle\") pod \"74372ec4-7697-478c-83d3-86a4742276e5\" (UID: \"74372ec4-7697-478c-83d3-86a4742276e5\") " Jan 24 11:43:59.095505 kubelet[2999]: I0124 11:43:59.093495 2999 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/74372ec4-7697-478c-83d3-86a4742276e5-whisker-backend-key-pair\") pod \"74372ec4-7697-478c-83d3-86a4742276e5\" (UID: \"74372ec4-7697-478c-83d3-86a4742276e5\") " Jan 24 11:43:59.117101 kubelet[2999]: I0124 11:43:59.116502 2999 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74372ec4-7697-478c-83d3-86a4742276e5-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "74372ec4-7697-478c-83d3-86a4742276e5" (UID: "74372ec4-7697-478c-83d3-86a4742276e5"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jan 24 11:43:59.186484 systemd[1]: var-lib-kubelet-pods-74372ec4\x2d7697\x2d478c\x2d83d3\x2d86a4742276e5-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Jan 24 11:43:59.212854 kubelet[2999]: I0124 11:43:59.212711 2999 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74372ec4-7697-478c-83d3-86a4742276e5-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "74372ec4-7697-478c-83d3-86a4742276e5" (UID: "74372ec4-7697-478c-83d3-86a4742276e5"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Jan 24 11:43:59.213044 systemd[1]: var-lib-kubelet-pods-74372ec4\x2d7697\x2d478c\x2d83d3\x2d86a4742276e5-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dxmcw2.mount: Deactivated successfully. Jan 24 11:43:59.214629 kubelet[2999]: I0124 11:43:59.214590 2999 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74372ec4-7697-478c-83d3-86a4742276e5-kube-api-access-xmcw2" (OuterVolumeSpecName: "kube-api-access-xmcw2") pod "74372ec4-7697-478c-83d3-86a4742276e5" (UID: "74372ec4-7697-478c-83d3-86a4742276e5"). InnerVolumeSpecName "kube-api-access-xmcw2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jan 24 11:43:59.215157 kubelet[2999]: I0124 11:43:59.214808 2999 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xmcw2\" (UniqueName: \"kubernetes.io/projected/74372ec4-7697-478c-83d3-86a4742276e5-kube-api-access-xmcw2\") on node \"localhost\" DevicePath \"\"" Jan 24 11:43:59.215157 kubelet[2999]: I0124 11:43:59.215068 2999 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74372ec4-7697-478c-83d3-86a4742276e5-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Jan 24 11:43:59.215157 kubelet[2999]: I0124 11:43:59.215084 2999 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/74372ec4-7697-478c-83d3-86a4742276e5-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Jan 24 11:43:59.221093 kubelet[2999]: E0124 11:43:59.221064 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:43:59.529130 systemd[1]: Removed slice kubepods-besteffort-pod74372ec4_7697_478c_83d3_86a4742276e5.slice - libcontainer container kubepods-besteffort-pod74372ec4_7697_478c_83d3_86a4742276e5.slice. Jan 24 11:43:59.995639 systemd[1]: Created slice kubepods-besteffort-podc2801bc8_955a_42fc_b0a7_f5018299bf2f.slice - libcontainer container kubepods-besteffort-podc2801bc8_955a_42fc_b0a7_f5018299bf2f.slice. Jan 24 11:44:00.132139 kubelet[2999]: I0124 11:44:00.131715 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4bkw\" (UniqueName: \"kubernetes.io/projected/c2801bc8-955a-42fc-b0a7-f5018299bf2f-kube-api-access-k4bkw\") pod \"whisker-65998ccc64-lw4d6\" (UID: \"c2801bc8-955a-42fc-b0a7-f5018299bf2f\") " pod="calico-system/whisker-65998ccc64-lw4d6" Jan 24 11:44:00.135497 kubelet[2999]: I0124 11:44:00.135344 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2801bc8-955a-42fc-b0a7-f5018299bf2f-whisker-ca-bundle\") pod \"whisker-65998ccc64-lw4d6\" (UID: \"c2801bc8-955a-42fc-b0a7-f5018299bf2f\") " pod="calico-system/whisker-65998ccc64-lw4d6" Jan 24 11:44:00.135497 kubelet[2999]: I0124 11:44:00.135410 2999 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/c2801bc8-955a-42fc-b0a7-f5018299bf2f-whisker-backend-key-pair\") pod \"whisker-65998ccc64-lw4d6\" (UID: \"c2801bc8-955a-42fc-b0a7-f5018299bf2f\") " pod="calico-system/whisker-65998ccc64-lw4d6" Jan 24 11:44:00.341699 containerd[1625]: time="2026-01-24T11:44:00.340809455Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-65998ccc64-lw4d6,Uid:c2801bc8-955a-42fc-b0a7-f5018299bf2f,Namespace:calico-system,Attempt:0,}" Jan 24 11:44:00.689046 kubelet[2999]: I0124 11:44:00.687835 2999 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74372ec4-7697-478c-83d3-86a4742276e5" path="/var/lib/kubelet/pods/74372ec4-7697-478c-83d3-86a4742276e5/volumes" Jan 24 11:44:00.826476 systemd-networkd[1501]: cali4c38279bfdc: Link UP Jan 24 11:44:00.860397 systemd-networkd[1501]: cali4c38279bfdc: Gained carrier Jan 24 11:44:00.986518 containerd[1625]: 2026-01-24 11:43:59.108 [INFO][4783] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 24 11:44:00.986518 containerd[1625]: 2026-01-24 11:43:59.269 [INFO][4783] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--7c778bb748--4gb25-eth0 goldmane-7c778bb748- calico-system 9f91eb4d-a73b-4530-82b4-682e0b2b659a 944 0 2026-01-24 11:42:49 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:7c778bb748 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-7c778bb748-4gb25 eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali4c38279bfdc [] [] }} ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Namespace="calico-system" Pod="goldmane-7c778bb748-4gb25" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--4gb25-" Jan 24 11:44:00.986518 containerd[1625]: 2026-01-24 11:43:59.273 [INFO][4783] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Namespace="calico-system" Pod="goldmane-7c778bb748-4gb25" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" Jan 24 11:44:00.986518 containerd[1625]: 2026-01-24 11:44:00.222 [INFO][4844] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" HandleID="k8s-pod-network.eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Workload="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.232 [INFO][4844] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" HandleID="k8s-pod-network.eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Workload="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00040c280), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-7c778bb748-4gb25", "timestamp":"2026-01-24 11:44:00.222396481 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.239 [INFO][4844] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.240 [INFO][4844] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.241 [INFO][4844] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.362 [INFO][4844] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" host="localhost" Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.467 [INFO][4844] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.530 [INFO][4844] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.538 [INFO][4844] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.554 [INFO][4844] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:00.992634 containerd[1625]: 2026-01-24 11:44:00.557 [INFO][4844] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" host="localhost" Jan 24 11:44:00.993493 containerd[1625]: 2026-01-24 11:44:00.595 [INFO][4844] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e Jan 24 11:44:00.993493 containerd[1625]: 2026-01-24 11:44:00.634 [INFO][4844] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" host="localhost" Jan 24 11:44:00.993493 containerd[1625]: 2026-01-24 11:44:00.676 [INFO][4844] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" host="localhost" Jan 24 11:44:00.993493 containerd[1625]: 2026-01-24 11:44:00.676 [INFO][4844] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" host="localhost" Jan 24 11:44:00.993493 containerd[1625]: 2026-01-24 11:44:00.683 [INFO][4844] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:44:00.993493 containerd[1625]: 2026-01-24 11:44:00.683 [INFO][4844] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" HandleID="k8s-pod-network.eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Workload="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" Jan 24 11:44:00.993735 containerd[1625]: 2026-01-24 11:44:00.706 [INFO][4783] cni-plugin/k8s.go 418: Populated endpoint ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Namespace="calico-system" Pod="goldmane-7c778bb748-4gb25" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--7c778bb748--4gb25-eth0", GenerateName:"goldmane-7c778bb748-", Namespace:"calico-system", SelfLink:"", UID:"9f91eb4d-a73b-4530-82b4-682e0b2b659a", ResourceVersion:"944", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7c778bb748", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-7c778bb748-4gb25", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4c38279bfdc", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:00.993735 containerd[1625]: 2026-01-24 11:44:00.708 [INFO][4783] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Namespace="calico-system" Pod="goldmane-7c778bb748-4gb25" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" Jan 24 11:44:00.994085 containerd[1625]: 2026-01-24 11:44:00.728 [INFO][4783] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4c38279bfdc ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Namespace="calico-system" Pod="goldmane-7c778bb748-4gb25" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" Jan 24 11:44:00.994085 containerd[1625]: 2026-01-24 11:44:00.863 [INFO][4783] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Namespace="calico-system" Pod="goldmane-7c778bb748-4gb25" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" Jan 24 11:44:00.994173 containerd[1625]: 2026-01-24 11:44:00.865 [INFO][4783] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Namespace="calico-system" Pod="goldmane-7c778bb748-4gb25" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--7c778bb748--4gb25-eth0", GenerateName:"goldmane-7c778bb748-", Namespace:"calico-system", SelfLink:"", UID:"9f91eb4d-a73b-4530-82b4-682e0b2b659a", ResourceVersion:"944", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7c778bb748", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e", Pod:"goldmane-7c778bb748-4gb25", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4c38279bfdc", MAC:"8e:7a:82:bc:ee:64", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:00.994526 containerd[1625]: 2026-01-24 11:44:00.968 [INFO][4783] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" Namespace="calico-system" Pod="goldmane-7c778bb748-4gb25" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--4gb25-eth0" Jan 24 11:44:01.118774 systemd-networkd[1501]: cali87147381262: Link UP Jan 24 11:44:01.121747 systemd-networkd[1501]: cali87147381262: Gained carrier Jan 24 11:44:01.242810 containerd[1625]: 2026-01-24 11:43:58.987 [INFO][4789] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 24 11:44:01.242810 containerd[1625]: 2026-01-24 11:43:59.271 [INFO][4789] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0 calico-kube-controllers-7dbccbb54b- calico-system c101de33-34a1-459f-bcfc-2378435cf35e 938 0 2026-01-24 11:42:53 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:7dbccbb54b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-7dbccbb54b-mvbw5 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali87147381262 [] [] }} ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Namespace="calico-system" Pod="calico-kube-controllers-7dbccbb54b-mvbw5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-" Jan 24 11:44:01.242810 containerd[1625]: 2026-01-24 11:43:59.275 [INFO][4789] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Namespace="calico-system" Pod="calico-kube-controllers-7dbccbb54b-mvbw5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" Jan 24 11:44:01.242810 containerd[1625]: 2026-01-24 11:44:00.230 [INFO][4843] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" HandleID="k8s-pod-network.416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Workload="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.241 [INFO][4843] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" HandleID="k8s-pod-network.416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Workload="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00038a2d0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-7dbccbb54b-mvbw5", "timestamp":"2026-01-24 11:44:00.230125276 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.241 [INFO][4843] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.678 [INFO][4843] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.678 [INFO][4843] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.735 [INFO][4843] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" host="localhost" Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.785 [INFO][4843] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.838 [INFO][4843] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.892 [INFO][4843] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.985 [INFO][4843] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:01.243523 containerd[1625]: 2026-01-24 11:44:00.985 [INFO][4843] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" host="localhost" Jan 24 11:44:01.244061 containerd[1625]: 2026-01-24 11:44:00.992 [INFO][4843] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515 Jan 24 11:44:01.244061 containerd[1625]: 2026-01-24 11:44:01.037 [INFO][4843] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" host="localhost" Jan 24 11:44:01.244061 containerd[1625]: 2026-01-24 11:44:01.080 [INFO][4843] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" host="localhost" Jan 24 11:44:01.244061 containerd[1625]: 2026-01-24 11:44:01.080 [INFO][4843] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" host="localhost" Jan 24 11:44:01.244061 containerd[1625]: 2026-01-24 11:44:01.081 [INFO][4843] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:44:01.244061 containerd[1625]: 2026-01-24 11:44:01.081 [INFO][4843] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" HandleID="k8s-pod-network.416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Workload="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" Jan 24 11:44:01.247469 containerd[1625]: 2026-01-24 11:44:01.097 [INFO][4789] cni-plugin/k8s.go 418: Populated endpoint ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Namespace="calico-system" Pod="calico-kube-controllers-7dbccbb54b-mvbw5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0", GenerateName:"calico-kube-controllers-7dbccbb54b-", Namespace:"calico-system", SelfLink:"", UID:"c101de33-34a1-459f-bcfc-2378435cf35e", ResourceVersion:"938", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7dbccbb54b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-7dbccbb54b-mvbw5", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali87147381262", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:01.247675 containerd[1625]: 2026-01-24 11:44:01.097 [INFO][4789] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Namespace="calico-system" Pod="calico-kube-controllers-7dbccbb54b-mvbw5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" Jan 24 11:44:01.247675 containerd[1625]: 2026-01-24 11:44:01.097 [INFO][4789] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali87147381262 ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Namespace="calico-system" Pod="calico-kube-controllers-7dbccbb54b-mvbw5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" Jan 24 11:44:01.247675 containerd[1625]: 2026-01-24 11:44:01.123 [INFO][4789] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Namespace="calico-system" Pod="calico-kube-controllers-7dbccbb54b-mvbw5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" Jan 24 11:44:01.247773 containerd[1625]: 2026-01-24 11:44:01.139 [INFO][4789] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Namespace="calico-system" Pod="calico-kube-controllers-7dbccbb54b-mvbw5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0", GenerateName:"calico-kube-controllers-7dbccbb54b-", Namespace:"calico-system", SelfLink:"", UID:"c101de33-34a1-459f-bcfc-2378435cf35e", ResourceVersion:"938", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7dbccbb54b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515", Pod:"calico-kube-controllers-7dbccbb54b-mvbw5", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali87147381262", MAC:"2a:8e:6a:2b:94:d3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:01.248030 containerd[1625]: 2026-01-24 11:44:01.189 [INFO][4789] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" Namespace="calico-system" Pod="calico-kube-controllers-7dbccbb54b-mvbw5" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7dbccbb54b--mvbw5-eth0" Jan 24 11:44:01.468628 containerd[1625]: time="2026-01-24T11:44:01.467538202Z" level=info msg="connecting to shim 416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515" address="unix:///run/containerd/s/0c1ac447f9e627901a618b659662298250c98c459907f662570c88017513926e" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:44:01.516013 containerd[1625]: time="2026-01-24T11:44:01.503751891Z" level=info msg="connecting to shim eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e" address="unix:///run/containerd/s/ee9ecba271020cb7dee42e95541baa665ba0e91c98072746a1e4e7bca2cd111d" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:44:01.532160 systemd-networkd[1501]: calicc8da4608d8: Link UP Jan 24 11:44:01.533711 systemd-networkd[1501]: calicc8da4608d8: Gained carrier Jan 24 11:44:01.645473 containerd[1625]: 2026-01-24 11:43:59.144 [INFO][4786] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 24 11:44:01.645473 containerd[1625]: 2026-01-24 11:43:59.266 [INFO][4786] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0 calico-apiserver-f58644cbc- calico-apiserver ec314a61-36d7-4af0-b9d6-11f6bfe86b8c 945 0 2026-01-24 11:42:44 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:f58644cbc projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-f58644cbc-hhctc eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calicc8da4608d8 [] [] }} ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-hhctc" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--hhctc-" Jan 24 11:44:01.645473 containerd[1625]: 2026-01-24 11:43:59.271 [INFO][4786] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-hhctc" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" Jan 24 11:44:01.645473 containerd[1625]: 2026-01-24 11:44:00.225 [INFO][4847] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" HandleID="k8s-pod-network.42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Workload="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:00.231 [INFO][4847] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" HandleID="k8s-pod-network.42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Workload="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d70e0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-f58644cbc-hhctc", "timestamp":"2026-01-24 11:44:00.225360816 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:00.231 [INFO][4847] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:01.081 [INFO][4847] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:01.082 [INFO][4847] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:01.123 [INFO][4847] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" host="localhost" Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:01.196 [INFO][4847] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:01.245 [INFO][4847] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:01.282 [INFO][4847] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:01.308 [INFO][4847] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:01.646755 containerd[1625]: 2026-01-24 11:44:01.308 [INFO][4847] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" host="localhost" Jan 24 11:44:01.649067 containerd[1625]: 2026-01-24 11:44:01.328 [INFO][4847] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4 Jan 24 11:44:01.649067 containerd[1625]: 2026-01-24 11:44:01.371 [INFO][4847] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" host="localhost" Jan 24 11:44:01.649067 containerd[1625]: 2026-01-24 11:44:01.423 [INFO][4847] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" host="localhost" Jan 24 11:44:01.649067 containerd[1625]: 2026-01-24 11:44:01.424 [INFO][4847] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" host="localhost" Jan 24 11:44:01.649067 containerd[1625]: 2026-01-24 11:44:01.445 [INFO][4847] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:44:01.649067 containerd[1625]: 2026-01-24 11:44:01.445 [INFO][4847] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" HandleID="k8s-pod-network.42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Workload="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" Jan 24 11:44:01.650016 containerd[1625]: 2026-01-24 11:44:01.515 [INFO][4786] cni-plugin/k8s.go 418: Populated endpoint ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-hhctc" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0", GenerateName:"calico-apiserver-f58644cbc-", Namespace:"calico-apiserver", SelfLink:"", UID:"ec314a61-36d7-4af0-b9d6-11f6bfe86b8c", ResourceVersion:"945", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f58644cbc", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-f58644cbc-hhctc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calicc8da4608d8", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:01.654646 containerd[1625]: 2026-01-24 11:44:01.517 [INFO][4786] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-hhctc" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" Jan 24 11:44:01.654646 containerd[1625]: 2026-01-24 11:44:01.517 [INFO][4786] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicc8da4608d8 ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-hhctc" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" Jan 24 11:44:01.654646 containerd[1625]: 2026-01-24 11:44:01.535 [INFO][4786] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-hhctc" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" Jan 24 11:44:01.655027 containerd[1625]: 2026-01-24 11:44:01.536 [INFO][4786] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-hhctc" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0", GenerateName:"calico-apiserver-f58644cbc-", Namespace:"calico-apiserver", SelfLink:"", UID:"ec314a61-36d7-4af0-b9d6-11f6bfe86b8c", ResourceVersion:"945", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f58644cbc", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4", Pod:"calico-apiserver-f58644cbc-hhctc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calicc8da4608d8", MAC:"62:cf:99:08:5c:91", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:01.655362 containerd[1625]: 2026-01-24 11:44:01.606 [INFO][4786] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-hhctc" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--hhctc-eth0" Jan 24 11:44:01.733702 systemd[1]: Started cri-containerd-416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515.scope - libcontainer container 416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515. Jan 24 11:44:01.739090 systemd[1]: Started cri-containerd-eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e.scope - libcontainer container eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e. Jan 24 11:44:01.832000 audit: BPF prog-id=173 op=LOAD Jan 24 11:44:01.833000 audit: BPF prog-id=174 op=LOAD Jan 24 11:44:01.833000 audit[4984]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000186238 a2=98 a3=0 items=0 ppid=4965 pid=4984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:01.833000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3431366332663763363466616130363138633036393964346232303365 Jan 24 11:44:01.834000 audit: BPF prog-id=174 op=UNLOAD Jan 24 11:44:01.834000 audit[4984]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4965 pid=4984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:01.834000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3431366332663763363466616130363138633036393964346232303365 Jan 24 11:44:01.834000 audit: BPF prog-id=175 op=LOAD Jan 24 11:44:01.834000 audit[4984]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=4965 pid=4984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:01.834000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3431366332663763363466616130363138633036393964346232303365 Jan 24 11:44:01.834000 audit: BPF prog-id=176 op=LOAD Jan 24 11:44:01.834000 audit[4984]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=4965 pid=4984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:01.834000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3431366332663763363466616130363138633036393964346232303365 Jan 24 11:44:01.834000 audit: BPF prog-id=176 op=UNLOAD Jan 24 11:44:01.834000 audit[4984]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=4965 pid=4984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:01.834000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3431366332663763363466616130363138633036393964346232303365 Jan 24 11:44:01.834000 audit: BPF prog-id=175 op=UNLOAD Jan 24 11:44:01.834000 audit[4984]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4965 pid=4984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:01.834000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3431366332663763363466616130363138633036393964346232303365 Jan 24 11:44:01.834000 audit: BPF prog-id=177 op=LOAD Jan 24 11:44:01.834000 audit[4984]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001866e8 a2=98 a3=0 items=0 ppid=4965 pid=4984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:01.834000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3431366332663763363466616130363138633036393964346232303365 Jan 24 11:44:01.852127 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:44:01.879574 containerd[1625]: time="2026-01-24T11:44:01.879523032Z" level=info msg="connecting to shim 42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4" address="unix:///run/containerd/s/1120667dcc272842c2cf3d702b37126740be5d68a5af9f0b23f11064c92bf248" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:44:01.905014 systemd-networkd[1501]: calib7400abaeb7: Link UP Jan 24 11:44:01.921845 systemd-networkd[1501]: calib7400abaeb7: Gained carrier Jan 24 11:44:02.023360 containerd[1625]: 2026-01-24 11:44:00.560 [INFO][4897] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 24 11:44:02.023360 containerd[1625]: 2026-01-24 11:44:00.640 [INFO][4897] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--65998ccc64--lw4d6-eth0 whisker-65998ccc64- calico-system c2801bc8-955a-42fc-b0a7-f5018299bf2f 1110 0 2026-01-24 11:43:59 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:65998ccc64 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-65998ccc64-lw4d6 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calib7400abaeb7 [] [] }} ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Namespace="calico-system" Pod="whisker-65998ccc64-lw4d6" WorkloadEndpoint="localhost-k8s-whisker--65998ccc64--lw4d6-" Jan 24 11:44:02.023360 containerd[1625]: 2026-01-24 11:44:00.646 [INFO][4897] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Namespace="calico-system" Pod="whisker-65998ccc64-lw4d6" WorkloadEndpoint="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" Jan 24 11:44:02.023360 containerd[1625]: 2026-01-24 11:44:00.936 [INFO][4916] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" HandleID="k8s-pod-network.bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Workload="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:00.937 [INFO][4916] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" HandleID="k8s-pod-network.bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Workload="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000590590), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-65998ccc64-lw4d6", "timestamp":"2026-01-24 11:44:00.936797495 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:00.937 [INFO][4916] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:01.451 [INFO][4916] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:01.471 [INFO][4916] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:01.528 [INFO][4916] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" host="localhost" Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:01.642 [INFO][4916] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:01.682 [INFO][4916] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:01.714 [INFO][4916] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:01.741 [INFO][4916] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:02.023823 containerd[1625]: 2026-01-24 11:44:01.753 [INFO][4916] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" host="localhost" Jan 24 11:44:02.024753 containerd[1625]: 2026-01-24 11:44:01.767 [INFO][4916] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3 Jan 24 11:44:02.024753 containerd[1625]: 2026-01-24 11:44:01.819 [INFO][4916] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" host="localhost" Jan 24 11:44:02.024753 containerd[1625]: 2026-01-24 11:44:01.864 [INFO][4916] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" host="localhost" Jan 24 11:44:02.024753 containerd[1625]: 2026-01-24 11:44:01.867 [INFO][4916] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" host="localhost" Jan 24 11:44:02.024753 containerd[1625]: 2026-01-24 11:44:01.867 [INFO][4916] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:44:02.024753 containerd[1625]: 2026-01-24 11:44:01.867 [INFO][4916] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" HandleID="k8s-pod-network.bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Workload="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" Jan 24 11:44:02.025041 containerd[1625]: 2026-01-24 11:44:01.894 [INFO][4897] cni-plugin/k8s.go 418: Populated endpoint ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Namespace="calico-system" Pod="whisker-65998ccc64-lw4d6" WorkloadEndpoint="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--65998ccc64--lw4d6-eth0", GenerateName:"whisker-65998ccc64-", Namespace:"calico-system", SelfLink:"", UID:"c2801bc8-955a-42fc-b0a7-f5018299bf2f", ResourceVersion:"1110", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 43, 59, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"65998ccc64", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-65998ccc64-lw4d6", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calib7400abaeb7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:02.025041 containerd[1625]: 2026-01-24 11:44:01.894 [INFO][4897] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Namespace="calico-system" Pod="whisker-65998ccc64-lw4d6" WorkloadEndpoint="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" Jan 24 11:44:02.025480 containerd[1625]: 2026-01-24 11:44:01.894 [INFO][4897] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib7400abaeb7 ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Namespace="calico-system" Pod="whisker-65998ccc64-lw4d6" WorkloadEndpoint="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" Jan 24 11:44:02.025480 containerd[1625]: 2026-01-24 11:44:01.918 [INFO][4897] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Namespace="calico-system" Pod="whisker-65998ccc64-lw4d6" WorkloadEndpoint="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" Jan 24 11:44:02.025563 containerd[1625]: 2026-01-24 11:44:01.918 [INFO][4897] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Namespace="calico-system" Pod="whisker-65998ccc64-lw4d6" WorkloadEndpoint="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--65998ccc64--lw4d6-eth0", GenerateName:"whisker-65998ccc64-", Namespace:"calico-system", SelfLink:"", UID:"c2801bc8-955a-42fc-b0a7-f5018299bf2f", ResourceVersion:"1110", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 43, 59, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"65998ccc64", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3", Pod:"whisker-65998ccc64-lw4d6", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calib7400abaeb7", MAC:"86:88:2d:52:32:1e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:02.025756 containerd[1625]: 2026-01-24 11:44:01.997 [INFO][4897] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" Namespace="calico-system" Pod="whisker-65998ccc64-lw4d6" WorkloadEndpoint="localhost-k8s-whisker--65998ccc64--lw4d6-eth0" Jan 24 11:44:02.046000 audit: BPF prog-id=178 op=LOAD Jan 24 11:44:02.053000 audit: BPF prog-id=179 op=LOAD Jan 24 11:44:02.053000 audit[4991]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186238 a2=98 a3=0 items=0 ppid=4967 pid=4991 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.053000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6562373761373264393364356637623132643461313930303032363334 Jan 24 11:44:02.056000 audit: BPF prog-id=179 op=UNLOAD Jan 24 11:44:02.056000 audit[4991]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4967 pid=4991 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.056000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6562373761373264393364356637623132643461313930303032363334 Jan 24 11:44:02.059000 audit: BPF prog-id=180 op=LOAD Jan 24 11:44:02.059000 audit[4991]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=4967 pid=4991 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.059000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6562373761373264393364356637623132643461313930303032363334 Jan 24 11:44:02.059000 audit: BPF prog-id=181 op=LOAD Jan 24 11:44:02.059000 audit[4991]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=4967 pid=4991 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.059000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6562373761373264393364356637623132643461313930303032363334 Jan 24 11:44:02.060000 audit: BPF prog-id=181 op=UNLOAD Jan 24 11:44:02.060000 audit[4991]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4967 pid=4991 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.060000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6562373761373264393364356637623132643461313930303032363334 Jan 24 11:44:02.062000 audit: BPF prog-id=180 op=UNLOAD Jan 24 11:44:02.062000 audit[4991]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4967 pid=4991 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.062000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6562373761373264393364356637623132643461313930303032363334 Jan 24 11:44:02.062000 audit: BPF prog-id=182 op=LOAD Jan 24 11:44:02.062000 audit[4991]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001866e8 a2=98 a3=0 items=0 ppid=4967 pid=4991 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.062000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6562373761373264393364356637623132643461313930303032363334 Jan 24 11:44:02.103139 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:44:02.124424 systemd[1]: Started cri-containerd-42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4.scope - libcontainer container 42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4. Jan 24 11:44:02.197019 kernel: kauditd_printk_skb: 49 callbacks suppressed Jan 24 11:44:02.197136 kernel: audit: type=1334 audit(1769255042.181:602): prog-id=183 op=LOAD Jan 24 11:44:02.181000 audit: BPF prog-id=183 op=LOAD Jan 24 11:44:02.196000 audit: BPF prog-id=184 op=LOAD Jan 24 11:44:02.196000 audit[5077]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.222677 containerd[1625]: time="2026-01-24T11:44:02.222122838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7dbccbb54b-mvbw5,Uid:c101de33-34a1-459f-bcfc-2378435cf35e,Namespace:calico-system,Attempt:0,} returns sandbox id \"416c2f7c64faa0618c0699d4b203ebde6da9e164294835dfd447f4cfe840a515\"" Jan 24 11:44:02.242545 kernel: audit: type=1334 audit(1769255042.196:603): prog-id=184 op=LOAD Jan 24 11:44:02.242712 kernel: audit: type=1300 audit(1769255042.196:603): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.245502 containerd[1625]: time="2026-01-24T11:44:02.245384167Z" level=info msg="connecting to shim bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3" address="unix:///run/containerd/s/fa2ba5e3b97f304b6b43da7b6dcbddcd1b19d884e0ac25133ddb65b8591bcb6f" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:44:02.245611 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:44:02.196000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.282104 containerd[1625]: time="2026-01-24T11:44:02.257527933Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 24 11:44:02.282447 kernel: audit: type=1327 audit(1769255042.196:603): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.196000 audit: BPF prog-id=184 op=UNLOAD Jan 24 11:44:02.314357 kernel: audit: type=1334 audit(1769255042.196:604): prog-id=184 op=UNLOAD Jan 24 11:44:02.314448 kernel: audit: type=1300 audit(1769255042.196:604): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.196000 audit[5077]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.382845 kernel: audit: type=1327 audit(1769255042.196:604): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.196000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.424257 kernel: audit: type=1334 audit(1769255042.196:605): prog-id=185 op=LOAD Jan 24 11:44:02.196000 audit: BPF prog-id=185 op=LOAD Jan 24 11:44:02.420325 systemd-networkd[1501]: cali87147381262: Gained IPv6LL Jan 24 11:44:02.196000 audit[5077]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.436610 kernel: audit: type=1300 audit(1769255042.196:605): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.436665 kernel: audit: type=1327 audit(1769255042.196:605): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.196000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.197000 audit: BPF prog-id=186 op=LOAD Jan 24 11:44:02.197000 audit[5077]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.197000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.197000 audit: BPF prog-id=186 op=UNLOAD Jan 24 11:44:02.197000 audit[5077]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.197000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.197000 audit: BPF prog-id=185 op=UNLOAD Jan 24 11:44:02.197000 audit[5077]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.197000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.197000 audit: BPF prog-id=187 op=LOAD Jan 24 11:44:02.197000 audit[5077]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=5044 pid=5077 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.197000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3432633863386165376331326666346131386432313235316130653939 Jan 24 11:44:02.477584 systemd-networkd[1501]: cali4c38279bfdc: Gained IPv6LL Jan 24 11:44:02.566977 containerd[1625]: time="2026-01-24T11:44:02.489532639Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:02.569441 containerd[1625]: time="2026-01-24T11:44:02.569382161Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 24 11:44:02.569672 containerd[1625]: time="2026-01-24T11:44:02.569643448Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:02.570348 kubelet[2999]: E0124 11:44:02.570144 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:44:02.574399 kubelet[2999]: E0124 11:44:02.573645 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:44:02.574399 kubelet[2999]: E0124 11:44:02.573845 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-kube-controllers start failed in pod calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:02.574399 kubelet[2999]: E0124 11:44:02.573983 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:44:02.620351 containerd[1625]: time="2026-01-24T11:44:02.606174582Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-4gb25,Uid:9f91eb4d-a73b-4530-82b4-682e0b2b659a,Namespace:calico-system,Attempt:0,} returns sandbox id \"eb77a72d93d5f7b12d4a190002634f92150a0e9a043a535f97ef55b79970f86e\"" Jan 24 11:44:02.654668 containerd[1625]: time="2026-01-24T11:44:02.654420889Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-hhctc,Uid:ec314a61-36d7-4af0-b9d6-11f6bfe86b8c,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"42c8c8ae7c12ff4a18d21251a0e99d0eaf24c72cb55d44aeb945554eaa646fb4\"" Jan 24 11:44:02.659125 containerd[1625]: time="2026-01-24T11:44:02.659092648Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 24 11:44:02.743433 containerd[1625]: time="2026-01-24T11:44:02.743111471Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:02.748699 containerd[1625]: time="2026-01-24T11:44:02.748630429Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 24 11:44:02.749040 containerd[1625]: time="2026-01-24T11:44:02.749006659Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:02.750500 systemd[1]: Started cri-containerd-bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3.scope - libcontainer container bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3. Jan 24 11:44:02.751097 kubelet[2999]: E0124 11:44:02.751027 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:44:02.751097 kubelet[2999]: E0124 11:44:02.751087 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:44:02.753487 kubelet[2999]: E0124 11:44:02.753440 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container goldmane start failed in pod goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:02.755097 kubelet[2999]: E0124 11:44:02.754696 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:44:02.757440 containerd[1625]: time="2026-01-24T11:44:02.757401972Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:44:02.844627 containerd[1625]: time="2026-01-24T11:44:02.844579729Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:02.848349 containerd[1625]: time="2026-01-24T11:44:02.848142692Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:44:02.848611 containerd[1625]: time="2026-01-24T11:44:02.848575670Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:02.854101 kubelet[2999]: E0124 11:44:02.853439 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:02.854668 kubelet[2999]: E0124 11:44:02.854630 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:02.856114 kubelet[2999]: E0124 11:44:02.855731 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:02.857406 kubelet[2999]: E0124 11:44:02.857365 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:44:02.897000 audit: BPF prog-id=188 op=LOAD Jan 24 11:44:02.912000 audit: BPF prog-id=189 op=LOAD Jan 24 11:44:02.912000 audit[5155]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000190238 a2=98 a3=0 items=0 ppid=5133 pid=5155 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.912000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633939396164303766613235653861363565656262653830313266 Jan 24 11:44:02.913000 audit: BPF prog-id=189 op=UNLOAD Jan 24 11:44:02.913000 audit[5155]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5133 pid=5155 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.913000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633939396164303766613235653861363565656262653830313266 Jan 24 11:44:02.916000 audit: BPF prog-id=190 op=LOAD Jan 24 11:44:02.916000 audit[5155]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000190488 a2=98 a3=0 items=0 ppid=5133 pid=5155 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.916000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633939396164303766613235653861363565656262653830313266 Jan 24 11:44:02.916000 audit: BPF prog-id=191 op=LOAD Jan 24 11:44:02.916000 audit[5155]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000190218 a2=98 a3=0 items=0 ppid=5133 pid=5155 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.916000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633939396164303766613235653861363565656262653830313266 Jan 24 11:44:02.916000 audit: BPF prog-id=191 op=UNLOAD Jan 24 11:44:02.916000 audit[5155]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5133 pid=5155 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.916000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633939396164303766613235653861363565656262653830313266 Jan 24 11:44:02.916000 audit: BPF prog-id=190 op=UNLOAD Jan 24 11:44:02.916000 audit[5155]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5133 pid=5155 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.916000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633939396164303766613235653861363565656262653830313266 Jan 24 11:44:02.916000 audit: BPF prog-id=192 op=LOAD Jan 24 11:44:02.916000 audit[5155]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001906e8 a2=98 a3=0 items=0 ppid=5133 pid=5155 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:02.916000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633939396164303766613235653861363565656262653830313266 Jan 24 11:44:02.923668 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:44:03.055054 systemd-networkd[1501]: calicc8da4608d8: Gained IPv6LL Jan 24 11:44:03.204307 containerd[1625]: time="2026-01-24T11:44:03.202980203Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-65998ccc64-lw4d6,Uid:c2801bc8-955a-42fc-b0a7-f5018299bf2f,Namespace:calico-system,Attempt:0,} returns sandbox id \"bcc999ad07fa25e8a65eebbe8012f2a9d7c35267a791be8f482eb39cbf5f20b3\"" Jan 24 11:44:03.215027 containerd[1625]: time="2026-01-24T11:44:03.214980926Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 24 11:44:03.308393 containerd[1625]: time="2026-01-24T11:44:03.307771337Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:03.311696 containerd[1625]: time="2026-01-24T11:44:03.310687192Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 24 11:44:03.311696 containerd[1625]: time="2026-01-24T11:44:03.311006867Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:03.311869 kubelet[2999]: E0124 11:44:03.311599 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:44:03.312703 kubelet[2999]: E0124 11:44:03.312396 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:44:03.312703 kubelet[2999]: E0124 11:44:03.312516 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker start failed in pod whisker-65998ccc64-lw4d6_calico-system(c2801bc8-955a-42fc-b0a7-f5018299bf2f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:03.316549 containerd[1625]: time="2026-01-24T11:44:03.316461285Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 24 11:44:03.342101 kubelet[2999]: E0124 11:44:03.340670 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:44:03.355107 kubelet[2999]: E0124 11:44:03.354680 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:44:03.355995 kubelet[2999]: E0124 11:44:03.355515 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:44:03.411169 containerd[1625]: time="2026-01-24T11:44:03.410713233Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:03.419566 containerd[1625]: time="2026-01-24T11:44:03.417022974Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 24 11:44:03.419566 containerd[1625]: time="2026-01-24T11:44:03.417372737Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:03.419801 kubelet[2999]: E0124 11:44:03.417845 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:44:03.419801 kubelet[2999]: E0124 11:44:03.418013 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:44:03.419801 kubelet[2999]: E0124 11:44:03.418128 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker-backend start failed in pod whisker-65998ccc64-lw4d6_calico-system(c2801bc8-955a-42fc-b0a7-f5018299bf2f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:03.419801 kubelet[2999]: E0124 11:44:03.418538 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:44:03.563000 audit[5250]: NETFILTER_CFG table=filter:119 family=2 entries=20 op=nft_register_rule pid=5250 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:03.563000 audit[5250]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7fff09db0370 a2=0 a3=7fff09db035c items=0 ppid=3157 pid=5250 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.563000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:03.578000 audit[5250]: NETFILTER_CFG table=nat:120 family=2 entries=14 op=nft_register_rule pid=5250 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:03.578000 audit[5250]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7fff09db0370 a2=0 a3=0 items=0 ppid=3157 pid=5250 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.578000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:03.759871 systemd-networkd[1501]: calib7400abaeb7: Gained IPv6LL Jan 24 11:44:03.805000 audit: BPF prog-id=193 op=LOAD Jan 24 11:44:03.805000 audit[5260]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffdb82f1030 a2=98 a3=1fffffffffffffff items=0 ppid=5135 pid=5260 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.805000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:44:03.806000 audit: BPF prog-id=193 op=UNLOAD Jan 24 11:44:03.806000 audit[5260]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffdb82f1000 a3=0 items=0 ppid=5135 pid=5260 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.806000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:44:03.806000 audit: BPF prog-id=194 op=LOAD Jan 24 11:44:03.806000 audit[5260]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffdb82f0f10 a2=94 a3=3 items=0 ppid=5135 pid=5260 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.806000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:44:03.806000 audit: BPF prog-id=194 op=UNLOAD Jan 24 11:44:03.806000 audit[5260]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffdb82f0f10 a2=94 a3=3 items=0 ppid=5135 pid=5260 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.806000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:44:03.806000 audit: BPF prog-id=195 op=LOAD Jan 24 11:44:03.806000 audit[5260]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffdb82f0f50 a2=94 a3=7ffdb82f1130 items=0 ppid=5135 pid=5260 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.806000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:44:03.806000 audit: BPF prog-id=195 op=UNLOAD Jan 24 11:44:03.806000 audit[5260]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffdb82f0f50 a2=94 a3=7ffdb82f1130 items=0 ppid=5135 pid=5260 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.806000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:44:03.829000 audit: BPF prog-id=196 op=LOAD Jan 24 11:44:03.829000 audit[5261]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffcc1484f40 a2=98 a3=3 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.829000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:03.831000 audit: BPF prog-id=196 op=UNLOAD Jan 24 11:44:03.831000 audit[5261]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffcc1484f10 a3=0 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.831000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:03.832000 audit: BPF prog-id=197 op=LOAD Jan 24 11:44:03.832000 audit[5261]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffcc1484d30 a2=94 a3=54428f items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.832000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:03.833000 audit: BPF prog-id=197 op=UNLOAD Jan 24 11:44:03.833000 audit[5261]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffcc1484d30 a2=94 a3=54428f items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.833000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:03.834000 audit: BPF prog-id=198 op=LOAD Jan 24 11:44:03.834000 audit[5261]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffcc1484d60 a2=94 a3=2 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.834000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:03.834000 audit: BPF prog-id=198 op=UNLOAD Jan 24 11:44:03.834000 audit[5261]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffcc1484d60 a2=0 a3=2 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:03.834000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.299000 audit: BPF prog-id=199 op=LOAD Jan 24 11:44:04.299000 audit[5261]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffcc1484c20 a2=94 a3=1 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.299000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.299000 audit: BPF prog-id=199 op=UNLOAD Jan 24 11:44:04.299000 audit[5261]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffcc1484c20 a2=94 a3=1 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.299000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.325000 audit: BPF prog-id=200 op=LOAD Jan 24 11:44:04.325000 audit[5261]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffcc1484c10 a2=94 a3=4 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.325000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.325000 audit: BPF prog-id=200 op=UNLOAD Jan 24 11:44:04.325000 audit[5261]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffcc1484c10 a2=0 a3=4 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.325000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.326000 audit: BPF prog-id=201 op=LOAD Jan 24 11:44:04.326000 audit[5261]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffcc1484a70 a2=94 a3=5 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.326000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.326000 audit: BPF prog-id=201 op=UNLOAD Jan 24 11:44:04.326000 audit[5261]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffcc1484a70 a2=0 a3=5 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.326000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.329000 audit: BPF prog-id=202 op=LOAD Jan 24 11:44:04.329000 audit[5261]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffcc1484c90 a2=94 a3=6 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.329000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.329000 audit: BPF prog-id=202 op=UNLOAD Jan 24 11:44:04.329000 audit[5261]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffcc1484c90 a2=0 a3=6 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.329000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.331000 audit: BPF prog-id=203 op=LOAD Jan 24 11:44:04.331000 audit[5261]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffcc1484440 a2=94 a3=88 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.331000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.333000 audit: BPF prog-id=204 op=LOAD Jan 24 11:44:04.333000 audit[5261]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7ffcc14842c0 a2=94 a3=2 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.333000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.333000 audit: BPF prog-id=204 op=UNLOAD Jan 24 11:44:04.333000 audit[5261]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7ffcc14842f0 a2=0 a3=7ffcc14843f0 items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.333000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.334000 audit: BPF prog-id=203 op=UNLOAD Jan 24 11:44:04.334000 audit[5261]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=1529ad10 a2=0 a3=cdb7817ce61624ad items=0 ppid=5135 pid=5261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.334000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:44:04.367590 kubelet[2999]: E0124 11:44:04.367451 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:44:04.374375 kubelet[2999]: E0124 11:44:04.374055 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:44:04.382149 kubelet[2999]: E0124 11:44:04.381469 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:44:04.389000 audit: BPF prog-id=205 op=LOAD Jan 24 11:44:04.389000 audit[5285]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe90101fd0 a2=98 a3=1999999999999999 items=0 ppid=5135 pid=5285 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.389000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:44:04.392000 audit: BPF prog-id=205 op=UNLOAD Jan 24 11:44:04.392000 audit[5285]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffe90101fa0 a3=0 items=0 ppid=5135 pid=5285 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.392000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:44:04.392000 audit: BPF prog-id=206 op=LOAD Jan 24 11:44:04.392000 audit[5285]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe90101eb0 a2=94 a3=ffff items=0 ppid=5135 pid=5285 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.392000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:44:04.392000 audit: BPF prog-id=206 op=UNLOAD Jan 24 11:44:04.392000 audit[5285]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffe90101eb0 a2=94 a3=ffff items=0 ppid=5135 pid=5285 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.392000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:44:04.392000 audit: BPF prog-id=207 op=LOAD Jan 24 11:44:04.392000 audit[5285]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe90101ef0 a2=94 a3=7ffe901020d0 items=0 ppid=5135 pid=5285 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.392000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:44:04.392000 audit: BPF prog-id=207 op=UNLOAD Jan 24 11:44:04.392000 audit[5285]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffe90101ef0 a2=94 a3=7ffe901020d0 items=0 ppid=5135 pid=5285 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.392000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:44:04.630000 audit[5301]: NETFILTER_CFG table=filter:121 family=2 entries=20 op=nft_register_rule pid=5301 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:04.630000 audit[5301]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffcb57da010 a2=0 a3=7ffcb57d9ffc items=0 ppid=3157 pid=5301 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.630000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:04.636000 audit[5301]: NETFILTER_CFG table=nat:122 family=2 entries=14 op=nft_register_rule pid=5301 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:04.636000 audit[5301]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffcb57da010 a2=0 a3=0 items=0 ppid=3157 pid=5301 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.636000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:04.822974 systemd-networkd[1501]: vxlan.calico: Link UP Jan 24 11:44:04.822994 systemd-networkd[1501]: vxlan.calico: Gained carrier Jan 24 11:44:04.944000 audit: BPF prog-id=208 op=LOAD Jan 24 11:44:04.944000 audit[5314]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd4a885910 a2=98 a3=0 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.944000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.944000 audit: BPF prog-id=208 op=UNLOAD Jan 24 11:44:04.944000 audit[5314]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffd4a8858e0 a3=0 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.944000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.945000 audit: BPF prog-id=209 op=LOAD Jan 24 11:44:04.945000 audit[5314]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd4a885720 a2=94 a3=54428f items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.945000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.945000 audit: BPF prog-id=209 op=UNLOAD Jan 24 11:44:04.945000 audit[5314]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd4a885720 a2=94 a3=54428f items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.945000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.945000 audit: BPF prog-id=210 op=LOAD Jan 24 11:44:04.945000 audit[5314]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd4a885750 a2=94 a3=2 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.945000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.945000 audit: BPF prog-id=210 op=UNLOAD Jan 24 11:44:04.945000 audit[5314]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd4a885750 a2=0 a3=2 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.945000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.945000 audit: BPF prog-id=211 op=LOAD Jan 24 11:44:04.945000 audit[5314]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffd4a885500 a2=94 a3=4 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.945000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.945000 audit: BPF prog-id=211 op=UNLOAD Jan 24 11:44:04.945000 audit[5314]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffd4a885500 a2=94 a3=4 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.945000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.946000 audit: BPF prog-id=212 op=LOAD Jan 24 11:44:04.946000 audit[5314]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffd4a885600 a2=94 a3=7ffd4a885780 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.946000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.946000 audit: BPF prog-id=212 op=UNLOAD Jan 24 11:44:04.946000 audit[5314]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffd4a885600 a2=0 a3=7ffd4a885780 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.946000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.947000 audit: BPF prog-id=213 op=LOAD Jan 24 11:44:04.947000 audit[5314]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffd4a884d30 a2=94 a3=2 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.947000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.947000 audit: BPF prog-id=213 op=UNLOAD Jan 24 11:44:04.947000 audit[5314]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffd4a884d30 a2=0 a3=2 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.947000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.947000 audit: BPF prog-id=214 op=LOAD Jan 24 11:44:04.947000 audit[5314]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffd4a884e30 a2=94 a3=30 items=0 ppid=5135 pid=5314 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.947000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:44:04.976000 audit: BPF prog-id=215 op=LOAD Jan 24 11:44:04.976000 audit[5318]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffff861e1b0 a2=98 a3=0 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.976000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:04.977000 audit: BPF prog-id=215 op=UNLOAD Jan 24 11:44:04.977000 audit[5318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffff861e180 a3=0 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.977000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:04.977000 audit: BPF prog-id=216 op=LOAD Jan 24 11:44:04.977000 audit[5318]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffff861dfa0 a2=94 a3=54428f items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.977000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:04.977000 audit: BPF prog-id=216 op=UNLOAD Jan 24 11:44:04.977000 audit[5318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffff861dfa0 a2=94 a3=54428f items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.977000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:04.977000 audit: BPF prog-id=217 op=LOAD Jan 24 11:44:04.977000 audit[5318]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffff861dfd0 a2=94 a3=2 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.977000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:04.978000 audit: BPF prog-id=217 op=UNLOAD Jan 24 11:44:04.978000 audit[5318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffff861dfd0 a2=0 a3=2 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:04.978000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.409000 audit: BPF prog-id=218 op=LOAD Jan 24 11:44:05.409000 audit[5318]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffff861de90 a2=94 a3=1 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.409000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.409000 audit: BPF prog-id=218 op=UNLOAD Jan 24 11:44:05.409000 audit[5318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffff861de90 a2=94 a3=1 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.409000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.424000 audit: BPF prog-id=219 op=LOAD Jan 24 11:44:05.424000 audit[5318]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffff861de80 a2=94 a3=4 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.424000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.425000 audit: BPF prog-id=219 op=UNLOAD Jan 24 11:44:05.425000 audit[5318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffff861de80 a2=0 a3=4 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.425000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.425000 audit: BPF prog-id=220 op=LOAD Jan 24 11:44:05.425000 audit[5318]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffff861dce0 a2=94 a3=5 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.425000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.426000 audit: BPF prog-id=220 op=UNLOAD Jan 24 11:44:05.426000 audit[5318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffff861dce0 a2=0 a3=5 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.426000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.426000 audit: BPF prog-id=221 op=LOAD Jan 24 11:44:05.426000 audit[5318]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffff861df00 a2=94 a3=6 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.426000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.426000 audit: BPF prog-id=221 op=UNLOAD Jan 24 11:44:05.426000 audit[5318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffff861df00 a2=0 a3=6 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.426000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.426000 audit: BPF prog-id=222 op=LOAD Jan 24 11:44:05.426000 audit[5318]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffff861d6b0 a2=94 a3=88 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.426000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.427000 audit: BPF prog-id=223 op=LOAD Jan 24 11:44:05.427000 audit[5318]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7ffff861d530 a2=94 a3=2 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.427000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.427000 audit: BPF prog-id=223 op=UNLOAD Jan 24 11:44:05.427000 audit[5318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7ffff861d560 a2=0 a3=7ffff861d660 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.427000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.428000 audit: BPF prog-id=222 op=UNLOAD Jan 24 11:44:05.428000 audit[5318]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=1947ad10 a2=0 a3=4e8f1e7b8e4cf880 items=0 ppid=5135 pid=5318 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.428000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:44:05.469000 audit: BPF prog-id=214 op=UNLOAD Jan 24 11:44:05.469000 audit[5135]: SYSCALL arch=c000003e syscall=263 success=yes exit=0 a0=ffffffffffffff9c a1=c000d2c240 a2=0 a3=0 items=0 ppid=5051 pid=5135 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="calico-node" exe="/usr/bin/calico-node" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.469000 audit: PROCTITLE proctitle=63616C69636F2D6E6F6465002D66656C6978 Jan 24 11:44:05.670084 containerd[1625]: time="2026-01-24T11:44:05.669672494Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:44:05.713000 audit[5350]: NETFILTER_CFG table=nat:123 family=2 entries=15 op=nft_register_chain pid=5350 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:44:05.713000 audit[5350]: SYSCALL arch=c000003e syscall=46 success=yes exit=5084 a0=3 a1=7ffd2193cd70 a2=0 a3=7ffd2193cd5c items=0 ppid=5135 pid=5350 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.713000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:44:05.741000 audit[5353]: NETFILTER_CFG table=mangle:124 family=2 entries=16 op=nft_register_chain pid=5353 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:44:05.741000 audit[5353]: SYSCALL arch=c000003e syscall=46 success=yes exit=6868 a0=3 a1=7ffec0d9e0e0 a2=0 a3=7ffec0d9e0cc items=0 ppid=5135 pid=5353 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.741000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:44:05.747000 audit[5348]: NETFILTER_CFG table=raw:125 family=2 entries=21 op=nft_register_chain pid=5348 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:44:05.747000 audit[5348]: SYSCALL arch=c000003e syscall=46 success=yes exit=8452 a0=3 a1=7ffd1d0a32a0 a2=0 a3=7ffd1d0a328c items=0 ppid=5135 pid=5348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.747000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:44:05.772000 audit[5365]: NETFILTER_CFG table=filter:126 family=2 entries=200 op=nft_register_chain pid=5365 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:44:05.772000 audit[5365]: SYSCALL arch=c000003e syscall=46 success=yes exit=117380 a0=3 a1=7ffdc3e811a0 a2=0 a3=7ffdc3e8118c items=0 ppid=5135 pid=5365 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:05.772000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:44:06.146752 systemd-networkd[1501]: calib1a17e01561: Link UP Jan 24 11:44:06.148749 systemd-networkd[1501]: calib1a17e01561: Gained carrier Jan 24 11:44:06.291164 containerd[1625]: 2026-01-24 11:44:05.861 [INFO][5351] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0 calico-apiserver-f58644cbc- calico-apiserver eda31195-c93b-4ab6-93d0-16b77a14ef46 942 0 2026-01-24 11:42:44 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:f58644cbc projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-f58644cbc-dvnfm eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calib1a17e01561 [] [] }} ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-dvnfm" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-" Jan 24 11:44:06.291164 containerd[1625]: 2026-01-24 11:44:05.861 [INFO][5351] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-dvnfm" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" Jan 24 11:44:06.291164 containerd[1625]: 2026-01-24 11:44:05.971 [INFO][5376] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" HandleID="k8s-pod-network.d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Workload="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:05.971 [INFO][5376] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" HandleID="k8s-pod-network.d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Workload="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002bd850), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-f58644cbc-dvnfm", "timestamp":"2026-01-24 11:44:05.971448904 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:05.971 [INFO][5376] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:05.972 [INFO][5376] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:05.972 [INFO][5376] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:05.993 [INFO][5376] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" host="localhost" Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:06.033 [INFO][5376] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:06.049 [INFO][5376] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:06.057 [INFO][5376] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:06.067 [INFO][5376] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:06.296082 containerd[1625]: 2026-01-24 11:44:06.067 [INFO][5376] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" host="localhost" Jan 24 11:44:06.307804 containerd[1625]: 2026-01-24 11:44:06.075 [INFO][5376] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6 Jan 24 11:44:06.307804 containerd[1625]: 2026-01-24 11:44:06.096 [INFO][5376] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" host="localhost" Jan 24 11:44:06.307804 containerd[1625]: 2026-01-24 11:44:06.121 [INFO][5376] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" host="localhost" Jan 24 11:44:06.307804 containerd[1625]: 2026-01-24 11:44:06.121 [INFO][5376] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" host="localhost" Jan 24 11:44:06.307804 containerd[1625]: 2026-01-24 11:44:06.121 [INFO][5376] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:44:06.307804 containerd[1625]: 2026-01-24 11:44:06.121 [INFO][5376] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" HandleID="k8s-pod-network.d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Workload="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" Jan 24 11:44:06.308160 containerd[1625]: 2026-01-24 11:44:06.131 [INFO][5351] cni-plugin/k8s.go 418: Populated endpoint ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-dvnfm" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0", GenerateName:"calico-apiserver-f58644cbc-", Namespace:"calico-apiserver", SelfLink:"", UID:"eda31195-c93b-4ab6-93d0-16b77a14ef46", ResourceVersion:"942", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f58644cbc", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-f58644cbc-dvnfm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calib1a17e01561", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:06.308589 containerd[1625]: 2026-01-24 11:44:06.131 [INFO][5351] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-dvnfm" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" Jan 24 11:44:06.308589 containerd[1625]: 2026-01-24 11:44:06.131 [INFO][5351] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib1a17e01561 ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-dvnfm" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" Jan 24 11:44:06.308589 containerd[1625]: 2026-01-24 11:44:06.150 [INFO][5351] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-dvnfm" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" Jan 24 11:44:06.308730 containerd[1625]: 2026-01-24 11:44:06.157 [INFO][5351] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-dvnfm" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0", GenerateName:"calico-apiserver-f58644cbc-", Namespace:"calico-apiserver", SelfLink:"", UID:"eda31195-c93b-4ab6-93d0-16b77a14ef46", ResourceVersion:"942", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f58644cbc", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6", Pod:"calico-apiserver-f58644cbc-dvnfm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calib1a17e01561", MAC:"62:48:66:a1:a2:f7", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:06.309032 containerd[1625]: 2026-01-24 11:44:06.268 [INFO][5351] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" Namespace="calico-apiserver" Pod="calico-apiserver-f58644cbc-dvnfm" WorkloadEndpoint="localhost-k8s-calico--apiserver--f58644cbc--dvnfm-eth0" Jan 24 11:44:06.329112 systemd-networkd[1501]: vxlan.calico: Gained IPv6LL Jan 24 11:44:06.462765 containerd[1625]: time="2026-01-24T11:44:06.462448913Z" level=info msg="connecting to shim d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6" address="unix:///run/containerd/s/072bc2dda5d200b6df54d76e56a93953f64d9b43bc2628c829027de72a66f8b7" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:44:06.614000 audit[5406]: NETFILTER_CFG table=filter:127 family=2 entries=55 op=nft_register_chain pid=5406 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:44:06.614000 audit[5406]: SYSCALL arch=c000003e syscall=46 success=yes exit=28304 a0=3 a1=7ffce987cdb0 a2=0 a3=7ffce987cd9c items=0 ppid=5135 pid=5406 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:06.614000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:44:06.671017 systemd[1]: Started cri-containerd-d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6.scope - libcontainer container d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6. Jan 24 11:44:06.685429 kubelet[2999]: E0124 11:44:06.685362 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:06.689883 containerd[1625]: time="2026-01-24T11:44:06.689707453Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,}" Jan 24 11:44:06.970000 audit: BPF prog-id=224 op=LOAD Jan 24 11:44:06.972000 audit: BPF prog-id=225 op=LOAD Jan 24 11:44:06.972000 audit[5412]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0238 a2=98 a3=0 items=0 ppid=5400 pid=5412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:06.972000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438393632343362653139306666303364386233346165396163343030 Jan 24 11:44:06.972000 audit: BPF prog-id=225 op=UNLOAD Jan 24 11:44:06.972000 audit[5412]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5400 pid=5412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:06.972000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438393632343362653139306666303364386233346165396163343030 Jan 24 11:44:06.972000 audit: BPF prog-id=226 op=LOAD Jan 24 11:44:06.972000 audit[5412]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=5400 pid=5412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:06.972000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438393632343362653139306666303364386233346165396163343030 Jan 24 11:44:06.974000 audit: BPF prog-id=227 op=LOAD Jan 24 11:44:06.974000 audit[5412]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=5400 pid=5412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:06.974000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438393632343362653139306666303364386233346165396163343030 Jan 24 11:44:06.974000 audit: BPF prog-id=227 op=UNLOAD Jan 24 11:44:06.974000 audit[5412]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5400 pid=5412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:06.974000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438393632343362653139306666303364386233346165396163343030 Jan 24 11:44:06.974000 audit: BPF prog-id=226 op=UNLOAD Jan 24 11:44:06.974000 audit[5412]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5400 pid=5412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:06.974000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438393632343362653139306666303364386233346165396163343030 Jan 24 11:44:06.974000 audit: BPF prog-id=228 op=LOAD Jan 24 11:44:06.974000 audit[5412]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=5400 pid=5412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:06.974000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438393632343362653139306666303364386233346165396163343030 Jan 24 11:44:06.986834 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:44:07.355095 containerd[1625]: time="2026-01-24T11:44:07.352679124Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f58644cbc-dvnfm,Uid:eda31195-c93b-4ab6-93d0-16b77a14ef46,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"d896243be190ff03d8b34ae9ac40034ee06a8c1fe7394749bfd9d0812daad6e6\"" Jan 24 11:44:07.361594 containerd[1625]: time="2026-01-24T11:44:07.361159002Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:44:07.390000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.26:22-10.0.0.1:46800 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:07.391391 systemd[1]: Started sshd@9-10.0.0.26:22-10.0.0.1:46800.service - OpenSSH per-connection server daemon (10.0.0.1:46800). Jan 24 11:44:07.401851 kernel: kauditd_printk_skb: 269 callbacks suppressed Jan 24 11:44:07.402077 kernel: audit: type=1130 audit(1769255047.390:697): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.26:22-10.0.0.1:46800 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:07.489176 containerd[1625]: time="2026-01-24T11:44:07.488511490Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:07.514674 containerd[1625]: time="2026-01-24T11:44:07.514622452Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:44:07.515319 containerd[1625]: time="2026-01-24T11:44:07.514872970Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:07.516867 kubelet[2999]: E0124 11:44:07.516785 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:07.516867 kubelet[2999]: E0124 11:44:07.516859 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:07.518443 kubelet[2999]: E0124 11:44:07.517050 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:07.518443 kubelet[2999]: E0124 11:44:07.517102 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:44:07.745886 systemd-networkd[1501]: cali130440c4247: Link UP Jan 24 11:44:07.766671 systemd-networkd[1501]: cali130440c4247: Gained carrier Jan 24 11:44:07.845151 sshd[5464]: Accepted publickey for core from 10.0.0.1 port 46800 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:07.843000 audit[5464]: USER_ACCT pid=5464 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:07.853174 containerd[1625]: 2026-01-24 11:44:07.049 [INFO][5427] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--66bc5c9577--zx6pd-eth0 coredns-66bc5c9577- kube-system ce8435a3-1751-4217-bed9-ea3dd2f1d3eb 939 0 2026-01-24 11:42:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-66bc5c9577-zx6pd eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali130440c4247 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Namespace="kube-system" Pod="coredns-66bc5c9577-zx6pd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--zx6pd-" Jan 24 11:44:07.853174 containerd[1625]: 2026-01-24 11:44:07.049 [INFO][5427] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Namespace="kube-system" Pod="coredns-66bc5c9577-zx6pd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" Jan 24 11:44:07.853174 containerd[1625]: 2026-01-24 11:44:07.324 [INFO][5449] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" HandleID="k8s-pod-network.e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Workload="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.327 [INFO][5449] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" HandleID="k8s-pod-network.e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Workload="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000bf590), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-66bc5c9577-zx6pd", "timestamp":"2026-01-24 11:44:07.32457592 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.328 [INFO][5449] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.328 [INFO][5449] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.329 [INFO][5449] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.368 [INFO][5449] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" host="localhost" Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.527 [INFO][5449] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.550 [INFO][5449] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.564 [INFO][5449] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.579 [INFO][5449] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:07.863375 containerd[1625]: 2026-01-24 11:44:07.579 [INFO][5449] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" host="localhost" Jan 24 11:44:07.864041 containerd[1625]: 2026-01-24 11:44:07.586 [INFO][5449] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176 Jan 24 11:44:07.864041 containerd[1625]: 2026-01-24 11:44:07.676 [INFO][5449] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" host="localhost" Jan 24 11:44:07.864041 containerd[1625]: 2026-01-24 11:44:07.710 [INFO][5449] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" host="localhost" Jan 24 11:44:07.864041 containerd[1625]: 2026-01-24 11:44:07.711 [INFO][5449] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" host="localhost" Jan 24 11:44:07.864041 containerd[1625]: 2026-01-24 11:44:07.711 [INFO][5449] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:44:07.864041 containerd[1625]: 2026-01-24 11:44:07.711 [INFO][5449] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" HandleID="k8s-pod-network.e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Workload="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" Jan 24 11:44:07.870628 containerd[1625]: 2026-01-24 11:44:07.725 [INFO][5427] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Namespace="kube-system" Pod="coredns-66bc5c9577-zx6pd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--zx6pd-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"ce8435a3-1751-4217-bed9-ea3dd2f1d3eb", ResourceVersion:"939", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-66bc5c9577-zx6pd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali130440c4247", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:07.870628 containerd[1625]: 2026-01-24 11:44:07.726 [INFO][5427] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Namespace="kube-system" Pod="coredns-66bc5c9577-zx6pd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" Jan 24 11:44:07.870628 containerd[1625]: 2026-01-24 11:44:07.726 [INFO][5427] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali130440c4247 ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Namespace="kube-system" Pod="coredns-66bc5c9577-zx6pd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" Jan 24 11:44:07.870628 containerd[1625]: 2026-01-24 11:44:07.780 [INFO][5427] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Namespace="kube-system" Pod="coredns-66bc5c9577-zx6pd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" Jan 24 11:44:07.870628 containerd[1625]: 2026-01-24 11:44:07.807 [INFO][5427] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Namespace="kube-system" Pod="coredns-66bc5c9577-zx6pd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--zx6pd-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"ce8435a3-1751-4217-bed9-ea3dd2f1d3eb", ResourceVersion:"939", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176", Pod:"coredns-66bc5c9577-zx6pd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali130440c4247", MAC:"da:7d:26:fc:86:6b", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:07.870628 containerd[1625]: 2026-01-24 11:44:07.837 [INFO][5427] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" Namespace="kube-system" Pod="coredns-66bc5c9577-zx6pd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--zx6pd-eth0" Jan 24 11:44:07.863000 audit[5464]: CRED_ACQ pid=5464 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:07.881825 sshd-session[5464]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:07.909558 kernel: audit: type=1101 audit(1769255047.843:698): pid=5464 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:07.909694 kernel: audit: type=1103 audit(1769255047.863:699): pid=5464 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:07.918706 systemd-logind[1579]: New session 11 of user core. Jan 24 11:44:07.863000 audit[5464]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd1e7c5770 a2=3 a3=0 items=0 ppid=1 pid=5464 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:07.961505 kernel: audit: type=1006 audit(1769255047.863:700): pid=5464 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=11 res=1 Jan 24 11:44:07.961660 kernel: audit: type=1300 audit(1769255047.863:700): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd1e7c5770 a2=3 a3=0 items=0 ppid=1 pid=5464 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:07.973701 kernel: audit: type=1327 audit(1769255047.863:700): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:07.863000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:07.991997 kernel: audit: type=1325 audit(1769255047.947:701): table=filter:128 family=2 entries=54 op=nft_register_chain pid=5477 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:44:07.947000 audit[5477]: NETFILTER_CFG table=filter:128 family=2 entries=54 op=nft_register_chain pid=5477 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:44:07.976592 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 24 11:44:07.947000 audit[5477]: SYSCALL arch=c000003e syscall=46 success=yes exit=26100 a0=3 a1=7fff01777f40 a2=0 a3=7fff01777f2c items=0 ppid=5135 pid=5477 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.033445 kernel: audit: type=1300 audit(1769255047.947:701): arch=c000003e syscall=46 success=yes exit=26100 a0=3 a1=7fff01777f40 a2=0 a3=7fff01777f2c items=0 ppid=5135 pid=5477 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.064382 kernel: audit: type=1327 audit(1769255047.947:701): proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:44:07.947000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:44:08.048352 systemd-networkd[1501]: calib1a17e01561: Gained IPv6LL Jan 24 11:44:08.064769 containerd[1625]: time="2026-01-24T11:44:08.039515892Z" level=info msg="connecting to shim e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176" address="unix:///run/containerd/s/7b16f01fa50612471ed6ed8630ada939bb50a823ebc9a1b9bff97a7220102c48" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:44:08.010000 audit[5464]: USER_START pid=5464 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:08.018000 audit[5492]: CRED_ACQ pid=5492 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:08.112591 kernel: audit: type=1105 audit(1769255048.010:702): pid=5464 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:08.214540 systemd[1]: Started cri-containerd-e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176.scope - libcontainer container e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176. Jan 24 11:44:08.289000 audit: BPF prog-id=229 op=LOAD Jan 24 11:44:08.292000 audit: BPF prog-id=230 op=LOAD Jan 24 11:44:08.292000 audit[5502]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=5486 pid=5502 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.292000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536393033613432623264633434353939313066643838646135366437 Jan 24 11:44:08.292000 audit: BPF prog-id=230 op=UNLOAD Jan 24 11:44:08.292000 audit[5502]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5486 pid=5502 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.292000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536393033613432623264633434353939313066643838646135366437 Jan 24 11:44:08.294000 audit: BPF prog-id=231 op=LOAD Jan 24 11:44:08.294000 audit[5502]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=5486 pid=5502 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.294000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536393033613432623264633434353939313066643838646135366437 Jan 24 11:44:08.294000 audit: BPF prog-id=232 op=LOAD Jan 24 11:44:08.294000 audit[5502]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=5486 pid=5502 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.294000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536393033613432623264633434353939313066643838646135366437 Jan 24 11:44:08.294000 audit: BPF prog-id=232 op=UNLOAD Jan 24 11:44:08.294000 audit[5502]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=5486 pid=5502 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.294000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536393033613432623264633434353939313066643838646135366437 Jan 24 11:44:08.294000 audit: BPF prog-id=231 op=UNLOAD Jan 24 11:44:08.294000 audit[5502]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5486 pid=5502 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.294000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536393033613432623264633434353939313066643838646135366437 Jan 24 11:44:08.294000 audit: BPF prog-id=233 op=LOAD Jan 24 11:44:08.294000 audit[5502]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=5486 pid=5502 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.294000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536393033613432623264633434353939313066643838646135366437 Jan 24 11:44:08.313989 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:44:08.467716 kubelet[2999]: E0124 11:44:08.466688 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:44:08.526510 containerd[1625]: time="2026-01-24T11:44:08.526126824Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-zx6pd,Uid:ce8435a3-1751-4217-bed9-ea3dd2f1d3eb,Namespace:kube-system,Attempt:0,} returns sandbox id \"e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176\"" Jan 24 11:44:08.530107 kubelet[2999]: E0124 11:44:08.530071 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:08.552494 containerd[1625]: time="2026-01-24T11:44:08.550989291Z" level=info msg="CreateContainer within sandbox \"e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 24 11:44:08.664089 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount304121798.mount: Deactivated successfully. Jan 24 11:44:08.675088 containerd[1625]: time="2026-01-24T11:44:08.674517532Z" level=info msg="Container b372b352625a3ee7594ac8a42cb8037761cbc350c92906d71d67c780482a915e: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:44:08.748993 containerd[1625]: time="2026-01-24T11:44:08.748694998Z" level=info msg="CreateContainer within sandbox \"e6903a42b2dc4459910fd88da56d7c38bca9bceb6c7c2548e64f9a775cc33176\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b372b352625a3ee7594ac8a42cb8037761cbc350c92906d71d67c780482a915e\"" Jan 24 11:44:08.751721 containerd[1625]: time="2026-01-24T11:44:08.750796836Z" level=info msg="StartContainer for \"b372b352625a3ee7594ac8a42cb8037761cbc350c92906d71d67c780482a915e\"" Jan 24 11:44:08.769710 containerd[1625]: time="2026-01-24T11:44:08.768127913Z" level=info msg="connecting to shim b372b352625a3ee7594ac8a42cb8037761cbc350c92906d71d67c780482a915e" address="unix:///run/containerd/s/7b16f01fa50612471ed6ed8630ada939bb50a823ebc9a1b9bff97a7220102c48" protocol=ttrpc version=3 Jan 24 11:44:08.774764 sshd[5492]: Connection closed by 10.0.0.1 port 46800 Jan 24 11:44:08.781349 sshd-session[5464]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:08.784000 audit[5464]: USER_END pid=5464 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:08.785000 audit[5464]: CRED_DISP pid=5464 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:08.798000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.26:22-10.0.0.1:46800 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:08.797126 systemd[1]: sshd@9-10.0.0.26:22-10.0.0.1:46800.service: Deactivated successfully. Jan 24 11:44:08.819788 systemd[1]: session-11.scope: Deactivated successfully. Jan 24 11:44:08.838839 systemd-logind[1579]: Session 11 logged out. Waiting for processes to exit. Jan 24 11:44:08.845046 systemd-logind[1579]: Removed session 11. Jan 24 11:44:08.852000 audit[5542]: NETFILTER_CFG table=filter:129 family=2 entries=20 op=nft_register_rule pid=5542 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:08.852000 audit[5542]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffcea1ae880 a2=0 a3=7ffcea1ae86c items=0 ppid=3157 pid=5542 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.852000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:08.867000 audit[5542]: NETFILTER_CFG table=nat:130 family=2 entries=14 op=nft_register_rule pid=5542 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:08.867000 audit[5542]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffcea1ae880 a2=0 a3=0 items=0 ppid=3157 pid=5542 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:08.867000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:08.878816 systemd-networkd[1501]: cali130440c4247: Gained IPv6LL Jan 24 11:44:08.889838 systemd[1]: Started cri-containerd-b372b352625a3ee7594ac8a42cb8037761cbc350c92906d71d67c780482a915e.scope - libcontainer container b372b352625a3ee7594ac8a42cb8037761cbc350c92906d71d67c780482a915e. Jan 24 11:44:09.034000 audit: BPF prog-id=234 op=LOAD Jan 24 11:44:09.037000 audit: BPF prog-id=235 op=LOAD Jan 24 11:44:09.037000 audit[5534]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=5486 pid=5534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:09.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6233373262333532363235613365653735393461633861343263623830 Jan 24 11:44:09.037000 audit: BPF prog-id=235 op=UNLOAD Jan 24 11:44:09.037000 audit[5534]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5486 pid=5534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:09.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6233373262333532363235613365653735393461633861343263623830 Jan 24 11:44:09.037000 audit: BPF prog-id=236 op=LOAD Jan 24 11:44:09.037000 audit[5534]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=5486 pid=5534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:09.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6233373262333532363235613365653735393461633861343263623830 Jan 24 11:44:09.037000 audit: BPF prog-id=237 op=LOAD Jan 24 11:44:09.037000 audit[5534]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=5486 pid=5534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:09.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6233373262333532363235613365653735393461633861343263623830 Jan 24 11:44:09.037000 audit: BPF prog-id=237 op=UNLOAD Jan 24 11:44:09.037000 audit[5534]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5486 pid=5534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:09.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6233373262333532363235613365653735393461633861343263623830 Jan 24 11:44:09.037000 audit: BPF prog-id=236 op=UNLOAD Jan 24 11:44:09.037000 audit[5534]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5486 pid=5534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:09.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6233373262333532363235613365653735393461633861343263623830 Jan 24 11:44:09.037000 audit: BPF prog-id=238 op=LOAD Jan 24 11:44:09.037000 audit[5534]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=5486 pid=5534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:09.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6233373262333532363235613365653735393461633861343263623830 Jan 24 11:44:09.192556 containerd[1625]: time="2026-01-24T11:44:09.192416551Z" level=info msg="StartContainer for \"b372b352625a3ee7594ac8a42cb8037761cbc350c92906d71d67c780482a915e\" returns successfully" Jan 24 11:44:09.479692 kubelet[2999]: E0124 11:44:09.479483 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:09.921000 audit[5571]: NETFILTER_CFG table=filter:131 family=2 entries=20 op=nft_register_rule pid=5571 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:09.921000 audit[5571]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffd173122c0 a2=0 a3=7ffd173122ac items=0 ppid=3157 pid=5571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:09.921000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:09.936000 audit[5571]: NETFILTER_CFG table=nat:132 family=2 entries=14 op=nft_register_rule pid=5571 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:09.936000 audit[5571]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffd173122c0 a2=0 a3=0 items=0 ppid=3157 pid=5571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:09.936000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:10.490484 kubelet[2999]: E0124 11:44:10.488673 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:10.557750 kubelet[2999]: I0124 11:44:10.557576 2999 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-zx6pd" podStartSLOduration=102.557550556 podStartE2EDuration="1m42.557550556s" podCreationTimestamp="2026-01-24 11:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:44:09.54360813 +0000 UTC m=+105.342835338" watchObservedRunningTime="2026-01-24 11:44:10.557550556 +0000 UTC m=+106.356777793" Jan 24 11:44:10.721789 kubelet[2999]: E0124 11:44:10.721568 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:10.728072 containerd[1625]: time="2026-01-24T11:44:10.726565907Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,}" Jan 24 11:44:11.070000 audit[5595]: NETFILTER_CFG table=filter:133 family=2 entries=17 op=nft_register_rule pid=5595 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:11.070000 audit[5595]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffeb86c3990 a2=0 a3=7ffeb86c397c items=0 ppid=3157 pid=5595 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:11.070000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:11.082000 audit[5595]: NETFILTER_CFG table=nat:134 family=2 entries=35 op=nft_register_chain pid=5595 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:11.082000 audit[5595]: SYSCALL arch=c000003e syscall=46 success=yes exit=14196 a0=3 a1=7ffeb86c3990 a2=0 a3=7ffeb86c397c items=0 ppid=3157 pid=5595 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:11.082000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:11.513527 kubelet[2999]: E0124 11:44:11.510824 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:11.829595 systemd-networkd[1501]: calic49c8d58e9a: Link UP Jan 24 11:44:11.834831 systemd-networkd[1501]: calic49c8d58e9a: Gained carrier Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.049 [INFO][5578] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--66bc5c9577--mlgbz-eth0 coredns-66bc5c9577- kube-system 6ef98b6a-4a5d-4999-ac24-f07f795beffa 934 0 2026-01-24 11:42:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-66bc5c9577-mlgbz eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calic49c8d58e9a [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Namespace="kube-system" Pod="coredns-66bc5c9577-mlgbz" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--mlgbz-" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.050 [INFO][5578] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Namespace="kube-system" Pod="coredns-66bc5c9577-mlgbz" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.327 [INFO][5597] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" HandleID="k8s-pod-network.7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Workload="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.329 [INFO][5597] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" HandleID="k8s-pod-network.7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Workload="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002219d0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-66bc5c9577-mlgbz", "timestamp":"2026-01-24 11:44:11.327835174 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.329 [INFO][5597] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.329 [INFO][5597] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.329 [INFO][5597] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.375 [INFO][5597] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" host="localhost" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.447 [INFO][5597] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.493 [INFO][5597] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.552 [INFO][5597] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.565 [INFO][5597] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.565 [INFO][5597] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" host="localhost" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.589 [INFO][5597] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04 Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.657 [INFO][5597] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" host="localhost" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.733 [INFO][5597] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" host="localhost" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.733 [INFO][5597] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" host="localhost" Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.733 [INFO][5597] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:44:11.952389 containerd[1625]: 2026-01-24 11:44:11.733 [INFO][5597] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" HandleID="k8s-pod-network.7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Workload="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" Jan 24 11:44:11.954824 containerd[1625]: 2026-01-24 11:44:11.753 [INFO][5578] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Namespace="kube-system" Pod="coredns-66bc5c9577-mlgbz" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--mlgbz-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"6ef98b6a-4a5d-4999-ac24-f07f795beffa", ResourceVersion:"934", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-66bc5c9577-mlgbz", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic49c8d58e9a", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:11.954824 containerd[1625]: 2026-01-24 11:44:11.753 [INFO][5578] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Namespace="kube-system" Pod="coredns-66bc5c9577-mlgbz" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" Jan 24 11:44:11.954824 containerd[1625]: 2026-01-24 11:44:11.753 [INFO][5578] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic49c8d58e9a ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Namespace="kube-system" Pod="coredns-66bc5c9577-mlgbz" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" Jan 24 11:44:11.954824 containerd[1625]: 2026-01-24 11:44:11.836 [INFO][5578] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Namespace="kube-system" Pod="coredns-66bc5c9577-mlgbz" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" Jan 24 11:44:11.954824 containerd[1625]: 2026-01-24 11:44:11.838 [INFO][5578] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Namespace="kube-system" Pod="coredns-66bc5c9577-mlgbz" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--mlgbz-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"6ef98b6a-4a5d-4999-ac24-f07f795beffa", ResourceVersion:"934", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04", Pod:"coredns-66bc5c9577-mlgbz", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic49c8d58e9a", MAC:"b2:d8:4d:9e:ea:6b", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:11.954824 containerd[1625]: 2026-01-24 11:44:11.907 [INFO][5578] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" Namespace="kube-system" Pod="coredns-66bc5c9577-mlgbz" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--mlgbz-eth0" Jan 24 11:44:12.050000 audit[5616]: NETFILTER_CFG table=filter:135 family=2 entries=54 op=nft_register_chain pid=5616 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:44:12.050000 audit[5616]: SYSCALL arch=c000003e syscall=46 success=yes exit=25556 a0=3 a1=7ffdaaabde60 a2=0 a3=7ffdaaabde4c items=0 ppid=5135 pid=5616 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.050000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:44:12.153166 containerd[1625]: time="2026-01-24T11:44:12.140890410Z" level=info msg="connecting to shim 7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04" address="unix:///run/containerd/s/519229ebab37689500163ff1ee63d70d42f9a7eca57d1691ba54aad77a2a50ec" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:44:12.400640 systemd[1]: Started cri-containerd-7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04.scope - libcontainer container 7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04. Jan 24 11:44:12.549000 audit: BPF prog-id=239 op=LOAD Jan 24 11:44:12.565986 kernel: kauditd_printk_skb: 69 callbacks suppressed Jan 24 11:44:12.566101 kernel: audit: type=1334 audit(1769255052.549:730): prog-id=239 op=LOAD Jan 24 11:44:12.576816 kernel: audit: type=1334 audit(1769255052.565:731): prog-id=240 op=LOAD Jan 24 11:44:12.565000 audit: BPF prog-id=240 op=LOAD Jan 24 11:44:12.565000 audit[5636]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8238 a2=98 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.630416 kernel: audit: type=1300 audit(1769255052.565:731): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8238 a2=98 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.565000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.637398 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:44:12.676345 kernel: audit: type=1327 audit(1769255052.565:731): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.676496 kernel: audit: type=1334 audit(1769255052.565:732): prog-id=240 op=UNLOAD Jan 24 11:44:12.565000 audit: BPF prog-id=240 op=UNLOAD Jan 24 11:44:12.565000 audit[5636]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.565000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.763859 kernel: audit: type=1300 audit(1769255052.565:732): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.764070 kernel: audit: type=1327 audit(1769255052.565:732): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.565000 audit: BPF prog-id=241 op=LOAD Jan 24 11:44:12.778379 kernel: audit: type=1334 audit(1769255052.565:733): prog-id=241 op=LOAD Jan 24 11:44:12.778543 kernel: audit: type=1300 audit(1769255052.565:733): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.565000 audit[5636]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.565000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.858163 kernel: audit: type=1327 audit(1769255052.565:733): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.565000 audit: BPF prog-id=242 op=LOAD Jan 24 11:44:12.565000 audit[5636]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a8218 a2=98 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.565000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.565000 audit: BPF prog-id=242 op=UNLOAD Jan 24 11:44:12.565000 audit[5636]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.565000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.565000 audit: BPF prog-id=241 op=UNLOAD Jan 24 11:44:12.565000 audit[5636]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.565000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.566000 audit: BPF prog-id=243 op=LOAD Jan 24 11:44:12.566000 audit[5636]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a86e8 a2=98 a3=0 items=0 ppid=5625 pid=5636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:12.566000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3763343166343839336639666230623137616233356137633862393231 Jan 24 11:44:12.961393 containerd[1625]: time="2026-01-24T11:44:12.957510726Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-mlgbz,Uid:6ef98b6a-4a5d-4999-ac24-f07f795beffa,Namespace:kube-system,Attempt:0,} returns sandbox id \"7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04\"" Jan 24 11:44:12.969714 kubelet[2999]: E0124 11:44:12.969522 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:12.998159 containerd[1625]: time="2026-01-24T11:44:12.994531437Z" level=info msg="CreateContainer within sandbox \"7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 24 11:44:13.084125 containerd[1625]: time="2026-01-24T11:44:13.081818178Z" level=info msg="Container d19c13820ff92671a2a52a028b7b1f2dc753e9274d7fad27a9881236d1472884: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:44:13.138418 containerd[1625]: time="2026-01-24T11:44:13.135407048Z" level=info msg="CreateContainer within sandbox \"7c41f4893f9fb0b17ab35a7c8b9210d358771c34110fa334d9ddb8a5a87a0a04\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d19c13820ff92671a2a52a028b7b1f2dc753e9274d7fad27a9881236d1472884\"" Jan 24 11:44:13.142162 containerd[1625]: time="2026-01-24T11:44:13.140694556Z" level=info msg="StartContainer for \"d19c13820ff92671a2a52a028b7b1f2dc753e9274d7fad27a9881236d1472884\"" Jan 24 11:44:13.150401 containerd[1625]: time="2026-01-24T11:44:13.148694369Z" level=info msg="connecting to shim d19c13820ff92671a2a52a028b7b1f2dc753e9274d7fad27a9881236d1472884" address="unix:///run/containerd/s/519229ebab37689500163ff1ee63d70d42f9a7eca57d1691ba54aad77a2a50ec" protocol=ttrpc version=3 Jan 24 11:44:13.231075 systemd[1]: Started cri-containerd-d19c13820ff92671a2a52a028b7b1f2dc753e9274d7fad27a9881236d1472884.scope - libcontainer container d19c13820ff92671a2a52a028b7b1f2dc753e9274d7fad27a9881236d1472884. Jan 24 11:44:13.288000 audit: BPF prog-id=244 op=LOAD Jan 24 11:44:13.292000 audit: BPF prog-id=245 op=LOAD Jan 24 11:44:13.292000 audit[5663]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186238 a2=98 a3=0 items=0 ppid=5625 pid=5663 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:13.292000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431396331333832306666393236373161326135326130323862376231 Jan 24 11:44:13.292000 audit: BPF prog-id=245 op=UNLOAD Jan 24 11:44:13.292000 audit[5663]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5625 pid=5663 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:13.292000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431396331333832306666393236373161326135326130323862376231 Jan 24 11:44:13.294000 audit: BPF prog-id=246 op=LOAD Jan 24 11:44:13.294000 audit[5663]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=5625 pid=5663 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:13.294000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431396331333832306666393236373161326135326130323862376231 Jan 24 11:44:13.300000 audit: BPF prog-id=247 op=LOAD Jan 24 11:44:13.300000 audit[5663]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=5625 pid=5663 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:13.300000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431396331333832306666393236373161326135326130323862376231 Jan 24 11:44:13.300000 audit: BPF prog-id=247 op=UNLOAD Jan 24 11:44:13.300000 audit[5663]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5625 pid=5663 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:13.300000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431396331333832306666393236373161326135326130323862376231 Jan 24 11:44:13.300000 audit: BPF prog-id=246 op=UNLOAD Jan 24 11:44:13.300000 audit[5663]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5625 pid=5663 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:13.300000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431396331333832306666393236373161326135326130323862376231 Jan 24 11:44:13.300000 audit: BPF prog-id=248 op=LOAD Jan 24 11:44:13.300000 audit[5663]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001866e8 a2=98 a3=0 items=0 ppid=5625 pid=5663 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:13.300000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431396331333832306666393236373161326135326130323862376231 Jan 24 11:44:13.455771 containerd[1625]: time="2026-01-24T11:44:13.455652567Z" level=info msg="StartContainer for \"d19c13820ff92671a2a52a028b7b1f2dc753e9274d7fad27a9881236d1472884\" returns successfully" Jan 24 11:44:13.550542 systemd-networkd[1501]: calic49c8d58e9a: Gained IPv6LL Jan 24 11:44:13.557632 kubelet[2999]: E0124 11:44:13.557457 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:13.668591 containerd[1625]: time="2026-01-24T11:44:13.668454207Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,}" Jan 24 11:44:13.736000 audit[5709]: NETFILTER_CFG table=filter:136 family=2 entries=14 op=nft_register_rule pid=5709 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:13.736000 audit[5709]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffea8102870 a2=0 a3=7ffea810285c items=0 ppid=3157 pid=5709 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:13.736000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:13.746000 audit[5709]: NETFILTER_CFG table=nat:137 family=2 entries=44 op=nft_register_rule pid=5709 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:13.746000 audit[5709]: SYSCALL arch=c000003e syscall=46 success=yes exit=14196 a0=3 a1=7ffea8102870 a2=0 a3=7ffea810285c items=0 ppid=3157 pid=5709 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:13.746000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:13.808000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.26:22-10.0.0.1:43980 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:13.809374 systemd[1]: Started sshd@10-10.0.0.26:22-10.0.0.1:43980.service - OpenSSH per-connection server daemon (10.0.0.1:43980). Jan 24 11:44:14.131000 audit[5714]: USER_ACCT pid=5714 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:14.134721 sshd[5714]: Accepted publickey for core from 10.0.0.1 port 43980 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:14.137000 audit[5714]: CRED_ACQ pid=5714 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:14.137000 audit[5714]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcb65dd680 a2=3 a3=0 items=0 ppid=1 pid=5714 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:14.137000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:14.140744 sshd-session[5714]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:14.156062 systemd-logind[1579]: New session 12 of user core. Jan 24 11:44:14.162004 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 24 11:44:14.181000 audit[5714]: USER_START pid=5714 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:14.189000 audit[5728]: CRED_ACQ pid=5728 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:14.304740 systemd-networkd[1501]: cali73398c139a2: Link UP Jan 24 11:44:14.305655 systemd-networkd[1501]: cali73398c139a2: Gained carrier Jan 24 11:44:14.411031 kubelet[2999]: I0124 11:44:14.402509 2999 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-mlgbz" podStartSLOduration=106.40248074 podStartE2EDuration="1m46.40248074s" podCreationTimestamp="2026-01-24 11:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:44:13.625037905 +0000 UTC m=+109.424265093" watchObservedRunningTime="2026-01-24 11:44:14.40248074 +0000 UTC m=+110.201707927" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:13.871 [INFO][5699] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--j5qxv-eth0 csi-node-driver- calico-system 9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd 810 0 2026-01-24 11:42:53 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:9d99788f7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-j5qxv eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali73398c139a2 [] [] }} ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Namespace="calico-system" Pod="csi-node-driver-j5qxv" WorkloadEndpoint="localhost-k8s-csi--node--driver--j5qxv-" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:13.871 [INFO][5699] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Namespace="calico-system" Pod="csi-node-driver-j5qxv" WorkloadEndpoint="localhost-k8s-csi--node--driver--j5qxv-eth0" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.005 [INFO][5718] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" HandleID="k8s-pod-network.9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Workload="localhost-k8s-csi--node--driver--j5qxv-eth0" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.007 [INFO][5718] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" HandleID="k8s-pod-network.9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Workload="localhost-k8s-csi--node--driver--j5qxv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004e530), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-j5qxv", "timestamp":"2026-01-24 11:44:14.005897268 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.007 [INFO][5718] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.007 [INFO][5718] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.008 [INFO][5718] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.050 [INFO][5718] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" host="localhost" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.085 [INFO][5718] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.120 [INFO][5718] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.133 [INFO][5718] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.152 [INFO][5718] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.152 [INFO][5718] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" host="localhost" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.162 [INFO][5718] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784 Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.190 [INFO][5718] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" host="localhost" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.242 [INFO][5718] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" host="localhost" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.248 [INFO][5718] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" host="localhost" Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.249 [INFO][5718] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:44:14.426405 containerd[1625]: 2026-01-24 11:44:14.249 [INFO][5718] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" HandleID="k8s-pod-network.9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Workload="localhost-k8s-csi--node--driver--j5qxv-eth0" Jan 24 11:44:14.433693 containerd[1625]: 2026-01-24 11:44:14.272 [INFO][5699] cni-plugin/k8s.go 418: Populated endpoint ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Namespace="calico-system" Pod="csi-node-driver-j5qxv" WorkloadEndpoint="localhost-k8s-csi--node--driver--j5qxv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--j5qxv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd", ResourceVersion:"810", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"9d99788f7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-j5qxv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali73398c139a2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:14.433693 containerd[1625]: 2026-01-24 11:44:14.272 [INFO][5699] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Namespace="calico-system" Pod="csi-node-driver-j5qxv" WorkloadEndpoint="localhost-k8s-csi--node--driver--j5qxv-eth0" Jan 24 11:44:14.433693 containerd[1625]: 2026-01-24 11:44:14.273 [INFO][5699] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali73398c139a2 ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Namespace="calico-system" Pod="csi-node-driver-j5qxv" WorkloadEndpoint="localhost-k8s-csi--node--driver--j5qxv-eth0" Jan 24 11:44:14.433693 containerd[1625]: 2026-01-24 11:44:14.313 [INFO][5699] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Namespace="calico-system" Pod="csi-node-driver-j5qxv" WorkloadEndpoint="localhost-k8s-csi--node--driver--j5qxv-eth0" Jan 24 11:44:14.433693 containerd[1625]: 2026-01-24 11:44:14.320 [INFO][5699] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Namespace="calico-system" Pod="csi-node-driver-j5qxv" WorkloadEndpoint="localhost-k8s-csi--node--driver--j5qxv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--j5qxv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd", ResourceVersion:"810", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 42, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"9d99788f7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784", Pod:"csi-node-driver-j5qxv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali73398c139a2", MAC:"16:8d:01:76:01:48", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:44:14.433693 containerd[1625]: 2026-01-24 11:44:14.414 [INFO][5699] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" Namespace="calico-system" Pod="csi-node-driver-j5qxv" WorkloadEndpoint="localhost-k8s-csi--node--driver--j5qxv-eth0" Jan 24 11:44:14.552000 audit[5747]: NETFILTER_CFG table=filter:138 family=2 entries=52 op=nft_register_chain pid=5747 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:44:14.552000 audit[5747]: SYSCALL arch=c000003e syscall=46 success=yes exit=24296 a0=3 a1=7fff45c4f980 a2=0 a3=7fff45c4f96c items=0 ppid=5135 pid=5747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:14.552000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:44:14.589471 kubelet[2999]: E0124 11:44:14.586731 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:14.677526 containerd[1625]: time="2026-01-24T11:44:14.677105075Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 24 11:44:14.760387 containerd[1625]: time="2026-01-24T11:44:14.759602729Z" level=info msg="connecting to shim 9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784" address="unix:///run/containerd/s/e1c2b68238664a96f0871fed44223372b47667ba22b9ee5e25570c0d4585c232" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:44:14.823385 sshd[5728]: Connection closed by 10.0.0.1 port 43980 Jan 24 11:44:14.821770 sshd-session[5714]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:14.824462 containerd[1625]: time="2026-01-24T11:44:14.818996184Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:14.833000 audit[5714]: USER_END pid=5714 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:14.833000 audit[5714]: CRED_DISP pid=5714 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:14.856712 systemd[1]: sshd@10-10.0.0.26:22-10.0.0.1:43980.service: Deactivated successfully. Jan 24 11:44:14.856000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.26:22-10.0.0.1:43980 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:14.866653 systemd[1]: session-12.scope: Deactivated successfully. Jan 24 11:44:14.871131 containerd[1625]: time="2026-01-24T11:44:14.871070304Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 24 11:44:14.871521 containerd[1625]: time="2026-01-24T11:44:14.871495206Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:14.872152 kubelet[2999]: E0124 11:44:14.871723 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:44:14.873418 kubelet[2999]: E0124 11:44:14.872730 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:44:14.873418 kubelet[2999]: E0124 11:44:14.872838 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-kube-controllers start failed in pod calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:14.873418 kubelet[2999]: E0124 11:44:14.872881 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:44:14.882553 systemd-logind[1579]: Session 12 logged out. Waiting for processes to exit. Jan 24 11:44:14.880000 audit[5774]: NETFILTER_CFG table=filter:139 family=2 entries=14 op=nft_register_rule pid=5774 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:14.880000 audit[5774]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fffb0d657e0 a2=0 a3=7fffb0d657cc items=0 ppid=3157 pid=5774 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:14.880000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:14.887456 systemd-logind[1579]: Removed session 12. Jan 24 11:44:14.943721 systemd[1]: Started cri-containerd-9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784.scope - libcontainer container 9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784. Jan 24 11:44:14.946000 audit[5774]: NETFILTER_CFG table=nat:140 family=2 entries=56 op=nft_register_chain pid=5774 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:44:14.946000 audit[5774]: SYSCALL arch=c000003e syscall=46 success=yes exit=19860 a0=3 a1=7fffb0d657e0 a2=0 a3=7fffb0d657cc items=0 ppid=3157 pid=5774 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:14.946000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:44:15.019000 audit: BPF prog-id=249 op=LOAD Jan 24 11:44:15.021000 audit: BPF prog-id=250 op=LOAD Jan 24 11:44:15.021000 audit[5773]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=5759 pid=5773 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:15.021000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3965643130346163353538333164303132303266356539393862363765 Jan 24 11:44:15.022000 audit: BPF prog-id=250 op=UNLOAD Jan 24 11:44:15.022000 audit[5773]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5759 pid=5773 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:15.022000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3965643130346163353538333164303132303266356539393862363765 Jan 24 11:44:15.023000 audit: BPF prog-id=251 op=LOAD Jan 24 11:44:15.023000 audit[5773]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=5759 pid=5773 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:15.023000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3965643130346163353538333164303132303266356539393862363765 Jan 24 11:44:15.024000 audit: BPF prog-id=252 op=LOAD Jan 24 11:44:15.024000 audit[5773]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=5759 pid=5773 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:15.024000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3965643130346163353538333164303132303266356539393862363765 Jan 24 11:44:15.025000 audit: BPF prog-id=252 op=UNLOAD Jan 24 11:44:15.025000 audit[5773]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=5759 pid=5773 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:15.025000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3965643130346163353538333164303132303266356539393862363765 Jan 24 11:44:15.025000 audit: BPF prog-id=251 op=UNLOAD Jan 24 11:44:15.025000 audit[5773]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5759 pid=5773 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:15.025000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3965643130346163353538333164303132303266356539393862363765 Jan 24 11:44:15.025000 audit: BPF prog-id=253 op=LOAD Jan 24 11:44:15.025000 audit[5773]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=5759 pid=5773 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:15.025000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3965643130346163353538333164303132303266356539393862363765 Jan 24 11:44:15.034423 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:44:15.146993 containerd[1625]: time="2026-01-24T11:44:15.146580238Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j5qxv,Uid:9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd,Namespace:calico-system,Attempt:0,} returns sandbox id \"9ed104ac55831d01202f5e998b67e111e5fae8255f07ae104a2888a47538b784\"" Jan 24 11:44:15.156773 containerd[1625]: time="2026-01-24T11:44:15.155143790Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 24 11:44:15.259635 containerd[1625]: time="2026-01-24T11:44:15.258527956Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:15.265087 containerd[1625]: time="2026-01-24T11:44:15.264859266Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 24 11:44:15.265087 containerd[1625]: time="2026-01-24T11:44:15.265006000Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:15.266097 kubelet[2999]: E0124 11:44:15.265868 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:44:15.266097 kubelet[2999]: E0124 11:44:15.266036 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:44:15.266400 kubelet[2999]: E0124 11:44:15.266142 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-csi start failed in pod csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:15.271899 containerd[1625]: time="2026-01-24T11:44:15.270569361Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 24 11:44:15.353430 containerd[1625]: time="2026-01-24T11:44:15.353101534Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:15.356726 containerd[1625]: time="2026-01-24T11:44:15.356435176Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 24 11:44:15.356726 containerd[1625]: time="2026-01-24T11:44:15.356583061Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:15.357699 kubelet[2999]: E0124 11:44:15.357519 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:44:15.357699 kubelet[2999]: E0124 11:44:15.357654 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:44:15.357820 kubelet[2999]: E0124 11:44:15.357755 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container csi-node-driver-registrar start failed in pod csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:15.357853 kubelet[2999]: E0124 11:44:15.357812 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:44:15.620642 kubelet[2999]: E0124 11:44:15.619354 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:15.629847 kubelet[2999]: E0124 11:44:15.629718 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:44:16.124735 systemd-networkd[1501]: cali73398c139a2: Gained IPv6LL Jan 24 11:44:16.639827 kubelet[2999]: E0124 11:44:16.634681 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:16.648896 kubelet[2999]: E0124 11:44:16.648045 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:44:16.668627 containerd[1625]: time="2026-01-24T11:44:16.668531539Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:44:16.777910 containerd[1625]: time="2026-01-24T11:44:16.777010259Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:16.781369 containerd[1625]: time="2026-01-24T11:44:16.781307990Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:44:16.781602 containerd[1625]: time="2026-01-24T11:44:16.781574096Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:16.784280 kubelet[2999]: E0124 11:44:16.783652 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:16.784280 kubelet[2999]: E0124 11:44:16.783717 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:16.784280 kubelet[2999]: E0124 11:44:16.784031 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:16.784280 kubelet[2999]: E0124 11:44:16.784088 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:44:16.784896 containerd[1625]: time="2026-01-24T11:44:16.784576929Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 24 11:44:16.974139 containerd[1625]: time="2026-01-24T11:44:16.974082670Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:16.979850 containerd[1625]: time="2026-01-24T11:44:16.979680856Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 24 11:44:16.979850 containerd[1625]: time="2026-01-24T11:44:16.979802393Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:16.981334 kubelet[2999]: E0124 11:44:16.981044 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:44:16.981334 kubelet[2999]: E0124 11:44:16.981152 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:44:16.983289 kubelet[2999]: E0124 11:44:16.981772 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker start failed in pod whisker-65998ccc64-lw4d6_calico-system(c2801bc8-955a-42fc-b0a7-f5018299bf2f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:16.984913 containerd[1625]: time="2026-01-24T11:44:16.984754114Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 24 11:44:17.081800 containerd[1625]: time="2026-01-24T11:44:17.081001835Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:17.086747 containerd[1625]: time="2026-01-24T11:44:17.086326634Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 24 11:44:17.086747 containerd[1625]: time="2026-01-24T11:44:17.086462658Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:17.090849 kubelet[2999]: E0124 11:44:17.090578 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:44:17.090849 kubelet[2999]: E0124 11:44:17.090655 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:44:17.090849 kubelet[2999]: E0124 11:44:17.090759 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker-backend start failed in pod whisker-65998ccc64-lw4d6_calico-system(c2801bc8-955a-42fc-b0a7-f5018299bf2f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:17.090849 kubelet[2999]: E0124 11:44:17.090813 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:44:17.664825 containerd[1625]: time="2026-01-24T11:44:17.664607589Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 24 11:44:17.749648 containerd[1625]: time="2026-01-24T11:44:17.749455293Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:17.752436 containerd[1625]: time="2026-01-24T11:44:17.752156778Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 24 11:44:17.752556 containerd[1625]: time="2026-01-24T11:44:17.752462749Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:17.754397 kubelet[2999]: E0124 11:44:17.752684 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:44:17.754397 kubelet[2999]: E0124 11:44:17.752822 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:44:17.754397 kubelet[2999]: E0124 11:44:17.753010 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container goldmane start failed in pod goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:17.754397 kubelet[2999]: E0124 11:44:17.753061 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:44:19.667782 containerd[1625]: time="2026-01-24T11:44:19.667473493Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:44:19.780912 containerd[1625]: time="2026-01-24T11:44:19.780592928Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:19.806486 containerd[1625]: time="2026-01-24T11:44:19.801444774Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:44:19.806486 containerd[1625]: time="2026-01-24T11:44:19.801798382Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:19.806734 kubelet[2999]: E0124 11:44:19.802419 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:19.806734 kubelet[2999]: E0124 11:44:19.802489 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:19.806734 kubelet[2999]: E0124 11:44:19.802591 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:19.806734 kubelet[2999]: E0124 11:44:19.802640 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:44:19.861714 systemd[1]: Started sshd@11-10.0.0.26:22-10.0.0.1:43982.service - OpenSSH per-connection server daemon (10.0.0.1:43982). Jan 24 11:44:19.860000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.26:22-10.0.0.1:43982 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:19.917673 kernel: kauditd_printk_skb: 82 callbacks suppressed Jan 24 11:44:19.917834 kernel: audit: type=1130 audit(1769255059.860:768): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.26:22-10.0.0.1:43982 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:20.178151 sshd[5813]: Accepted publickey for core from 10.0.0.1 port 43982 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:20.174000 audit[5813]: USER_ACCT pid=5813 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.235132 kernel: audit: type=1101 audit(1769255060.174:769): pid=5813 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.187439 sshd-session[5813]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:20.180000 audit[5813]: CRED_ACQ pid=5813 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.247599 systemd-logind[1579]: New session 13 of user core. Jan 24 11:44:20.300909 kernel: audit: type=1103 audit(1769255060.180:770): pid=5813 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.301497 kernel: audit: type=1006 audit(1769255060.180:771): pid=5813 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=13 res=1 Jan 24 11:44:20.180000 audit[5813]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe6354f560 a2=3 a3=0 items=0 ppid=1 pid=5813 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:20.374693 kernel: audit: type=1300 audit(1769255060.180:771): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe6354f560 a2=3 a3=0 items=0 ppid=1 pid=5813 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:20.374914 kernel: audit: type=1327 audit(1769255060.180:771): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:20.180000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:20.375748 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 24 11:44:20.429000 audit[5813]: USER_START pid=5813 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.473346 kernel: audit: type=1105 audit(1769255060.429:772): pid=5813 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.473481 kernel: audit: type=1103 audit(1769255060.437:773): pid=5817 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.437000 audit[5817]: CRED_ACQ pid=5817 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.885377 sshd[5817]: Connection closed by 10.0.0.1 port 43982 Jan 24 11:44:20.887049 sshd-session[5813]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:20.889000 audit[5813]: USER_END pid=5813 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.924535 systemd[1]: sshd@11-10.0.0.26:22-10.0.0.1:43982.service: Deactivated successfully. Jan 24 11:44:20.936613 systemd[1]: session-13.scope: Deactivated successfully. Jan 24 11:44:20.946571 systemd-logind[1579]: Session 13 logged out. Waiting for processes to exit. Jan 24 11:44:20.955115 systemd-logind[1579]: Removed session 13. Jan 24 11:44:20.889000 audit[5813]: CRED_DISP pid=5813 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:21.020450 kernel: audit: type=1106 audit(1769255060.889:774): pid=5813 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:21.020575 kernel: audit: type=1104 audit(1769255060.889:775): pid=5813 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:20.921000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.26:22-10.0.0.1:43982 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:25.938438 systemd[1]: Started sshd@12-10.0.0.26:22-10.0.0.1:33474.service - OpenSSH per-connection server daemon (10.0.0.1:33474). Jan 24 11:44:25.937000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.26:22-10.0.0.1:33474 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:25.956335 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:44:25.956574 kernel: audit: type=1130 audit(1769255065.937:777): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.26:22-10.0.0.1:33474 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:26.181000 audit[5843]: USER_ACCT pid=5843 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.182866 sshd[5843]: Accepted publickey for core from 10.0.0.1 port 33474 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:26.191825 sshd-session[5843]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:26.228908 systemd-logind[1579]: New session 14 of user core. Jan 24 11:44:26.259685 kernel: audit: type=1101 audit(1769255066.181:778): pid=5843 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.185000 audit[5843]: CRED_ACQ pid=5843 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.339524 kernel: audit: type=1103 audit(1769255066.185:779): pid=5843 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.339742 kernel: audit: type=1006 audit(1769255066.185:780): pid=5843 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=14 res=1 Jan 24 11:44:26.341411 kernel: audit: type=1300 audit(1769255066.185:780): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffde51701f0 a2=3 a3=0 items=0 ppid=1 pid=5843 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:26.185000 audit[5843]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffde51701f0 a2=3 a3=0 items=0 ppid=1 pid=5843 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:26.351624 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 24 11:44:26.185000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:26.407705 kernel: audit: type=1327 audit(1769255066.185:780): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:26.366000 audit[5843]: USER_START pid=5843 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.449326 kernel: audit: type=1105 audit(1769255066.366:781): pid=5843 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.374000 audit[5847]: CRED_ACQ pid=5847 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.521454 kernel: audit: type=1103 audit(1769255066.374:782): pid=5847 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.885841 sshd[5847]: Connection closed by 10.0.0.1 port 33474 Jan 24 11:44:26.886608 sshd-session[5843]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:26.887000 audit[5843]: USER_END pid=5843 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.927383 kernel: audit: type=1106 audit(1769255066.887:783): pid=5843 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.928944 systemd[1]: sshd@12-10.0.0.26:22-10.0.0.1:33474.service: Deactivated successfully. Jan 24 11:44:26.934508 systemd[1]: session-14.scope: Deactivated successfully. Jan 24 11:44:26.892000 audit[5843]: CRED_DISP pid=5843 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:26.941023 systemd-logind[1579]: Session 14 logged out. Waiting for processes to exit. Jan 24 11:44:26.944651 systemd-logind[1579]: Removed session 14. Jan 24 11:44:26.928000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.26:22-10.0.0.1:33474 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:26.958427 kernel: audit: type=1104 audit(1769255066.892:784): pid=5843 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:27.669676 kubelet[2999]: E0124 11:44:27.669505 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:44:27.669676 kubelet[2999]: E0124 11:44:27.669505 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:44:28.676168 kubelet[2999]: E0124 11:44:28.675797 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:44:29.671252 kubelet[2999]: E0124 11:44:29.669499 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:44:30.493085 kubelet[2999]: E0124 11:44:30.492834 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:31.664683 containerd[1625]: time="2026-01-24T11:44:31.664578823Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 24 11:44:31.775071 containerd[1625]: time="2026-01-24T11:44:31.772619280Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:31.776509 containerd[1625]: time="2026-01-24T11:44:31.776349165Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 24 11:44:31.776509 containerd[1625]: time="2026-01-24T11:44:31.776467174Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:31.776869 kubelet[2999]: E0124 11:44:31.776820 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:44:31.778919 kubelet[2999]: E0124 11:44:31.778548 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:44:31.778919 kubelet[2999]: E0124 11:44:31.783450 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-csi start failed in pod csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:31.789334 containerd[1625]: time="2026-01-24T11:44:31.785742498Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 24 11:44:31.896620 containerd[1625]: time="2026-01-24T11:44:31.895562174Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:31.939311 containerd[1625]: time="2026-01-24T11:44:31.938512437Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 24 11:44:31.939311 containerd[1625]: time="2026-01-24T11:44:31.938652377Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:31.940419 kubelet[2999]: E0124 11:44:31.940075 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:44:31.940419 kubelet[2999]: E0124 11:44:31.940138 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:44:31.947128 kubelet[2999]: E0124 11:44:31.946639 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container csi-node-driver-registrar start failed in pod csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:31.947128 kubelet[2999]: E0124 11:44:31.946769 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:44:31.946000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.26:22-10.0.0.1:33486 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:31.947394 systemd[1]: Started sshd@13-10.0.0.26:22-10.0.0.1:33486.service - OpenSSH per-connection server daemon (10.0.0.1:33486). Jan 24 11:44:31.960834 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:44:31.962406 kernel: audit: type=1130 audit(1769255071.946:786): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.26:22-10.0.0.1:33486 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:32.193000 audit[5892]: USER_ACCT pid=5892 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.196837 sshd[5892]: Accepted publickey for core from 10.0.0.1 port 33486 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:32.210627 sshd-session[5892]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:32.275448 kernel: audit: type=1101 audit(1769255072.193:787): pid=5892 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.275728 kernel: audit: type=1103 audit(1769255072.201:788): pid=5892 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.201000 audit[5892]: CRED_ACQ pid=5892 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.267832 systemd-logind[1579]: New session 15 of user core. Jan 24 11:44:32.203000 audit[5892]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffd35af560 a2=3 a3=0 items=0 ppid=1 pid=5892 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:32.346371 kernel: audit: type=1006 audit(1769255072.203:789): pid=5892 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=15 res=1 Jan 24 11:44:32.346467 kernel: audit: type=1300 audit(1769255072.203:789): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffd35af560 a2=3 a3=0 items=0 ppid=1 pid=5892 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:32.346517 kernel: audit: type=1327 audit(1769255072.203:789): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:32.203000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:32.347782 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 24 11:44:32.356000 audit[5892]: USER_START pid=5892 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.368000 audit[5897]: CRED_ACQ pid=5897 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.438764 kernel: audit: type=1105 audit(1769255072.356:790): pid=5892 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.438877 kernel: audit: type=1103 audit(1769255072.368:791): pid=5897 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.887125 sshd[5897]: Connection closed by 10.0.0.1 port 33486 Jan 24 11:44:32.888713 sshd-session[5892]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:32.893000 audit[5892]: USER_END pid=5892 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.955574 kernel: audit: type=1106 audit(1769255072.893:792): pid=5892 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.904000 audit[5892]: CRED_DISP pid=5892 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.958797 systemd[1]: sshd@13-10.0.0.26:22-10.0.0.1:33486.service: Deactivated successfully. Jan 24 11:44:32.965948 systemd[1]: session-15.scope: Deactivated successfully. Jan 24 11:44:32.978069 systemd-logind[1579]: Session 15 logged out. Waiting for processes to exit. Jan 24 11:44:32.984633 kernel: audit: type=1104 audit(1769255072.904:793): pid=5892 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:32.956000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.26:22-10.0.0.1:33486 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:32.989528 systemd-logind[1579]: Removed session 15. Jan 24 11:44:34.680314 kubelet[2999]: E0124 11:44:34.675559 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:44:37.912000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.26:22-10.0.0.1:58678 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:37.913123 systemd[1]: Started sshd@14-10.0.0.26:22-10.0.0.1:58678.service - OpenSSH per-connection server daemon (10.0.0.1:58678). Jan 24 11:44:37.920973 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:44:37.921334 kernel: audit: type=1130 audit(1769255077.912:795): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.26:22-10.0.0.1:58678 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:38.081000 audit[5919]: USER_ACCT pid=5919 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.084903 sshd[5919]: Accepted publickey for core from 10.0.0.1 port 58678 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:38.086496 sshd-session[5919]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:38.097957 systemd-logind[1579]: New session 16 of user core. Jan 24 11:44:38.125412 kernel: audit: type=1101 audit(1769255078.081:796): pid=5919 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.125520 kernel: audit: type=1103 audit(1769255078.083:797): pid=5919 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.083000 audit[5919]: CRED_ACQ pid=5919 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.149606 kernel: audit: type=1006 audit(1769255078.083:798): pid=5919 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=16 res=1 Jan 24 11:44:38.083000 audit[5919]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffbee0acb0 a2=3 a3=0 items=0 ppid=1 pid=5919 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:38.195339 kernel: audit: type=1300 audit(1769255078.083:798): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffbee0acb0 a2=3 a3=0 items=0 ppid=1 pid=5919 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:38.195520 kernel: audit: type=1327 audit(1769255078.083:798): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:38.083000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:38.198375 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 24 11:44:38.226000 audit[5919]: USER_START pid=5919 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.267472 kernel: audit: type=1105 audit(1769255078.226:799): pid=5919 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.267635 kernel: audit: type=1103 audit(1769255078.235:800): pid=5923 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.235000 audit[5923]: CRED_ACQ pid=5923 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.658599 sshd[5923]: Connection closed by 10.0.0.1 port 58678 Jan 24 11:44:38.661500 sshd-session[5919]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:38.668000 audit[5919]: USER_END pid=5919 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.675576 containerd[1625]: time="2026-01-24T11:44:38.674560417Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 24 11:44:38.718529 kernel: audit: type=1106 audit(1769255078.668:801): pid=5919 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.718604 systemd[1]: sshd@14-10.0.0.26:22-10.0.0.1:58678.service: Deactivated successfully. Jan 24 11:44:38.722904 systemd[1]: session-16.scope: Deactivated successfully. Jan 24 11:44:38.668000 audit[5919]: CRED_DISP pid=5919 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.727367 systemd-logind[1579]: Session 16 logged out. Waiting for processes to exit. Jan 24 11:44:38.731699 systemd-logind[1579]: Removed session 16. Jan 24 11:44:38.754457 kernel: audit: type=1104 audit(1769255078.668:802): pid=5919 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:38.718000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.26:22-10.0.0.1:58678 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:38.818322 containerd[1625]: time="2026-01-24T11:44:38.818083903Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:38.829949 containerd[1625]: time="2026-01-24T11:44:38.829764582Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 24 11:44:38.829949 containerd[1625]: time="2026-01-24T11:44:38.829896699Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:38.832355 kubelet[2999]: E0124 11:44:38.830948 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:44:38.833959 kubelet[2999]: E0124 11:44:38.832977 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:44:38.833959 kubelet[2999]: E0124 11:44:38.833463 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-kube-controllers start failed in pod calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:38.833959 kubelet[2999]: E0124 11:44:38.833513 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:44:39.668618 containerd[1625]: time="2026-01-24T11:44:39.668562858Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:44:39.765734 containerd[1625]: time="2026-01-24T11:44:39.765668987Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:39.775961 containerd[1625]: time="2026-01-24T11:44:39.775747431Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:44:39.775961 containerd[1625]: time="2026-01-24T11:44:39.775893153Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:39.779156 kubelet[2999]: E0124 11:44:39.778686 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:39.779156 kubelet[2999]: E0124 11:44:39.778757 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:39.779156 kubelet[2999]: E0124 11:44:39.778873 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:39.779156 kubelet[2999]: E0124 11:44:39.779317 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:44:42.693372 containerd[1625]: time="2026-01-24T11:44:42.688346361Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 24 11:44:42.697655 kubelet[2999]: E0124 11:44:42.690796 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:44:42.814457 containerd[1625]: time="2026-01-24T11:44:42.808824152Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:42.814457 containerd[1625]: time="2026-01-24T11:44:42.813909164Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 24 11:44:42.814676 kubelet[2999]: E0124 11:44:42.814425 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:44:42.814676 kubelet[2999]: E0124 11:44:42.814480 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:44:42.815689 kubelet[2999]: E0124 11:44:42.815074 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container goldmane start failed in pod goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:42.815689 kubelet[2999]: E0124 11:44:42.815442 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:44:42.815798 containerd[1625]: time="2026-01-24T11:44:42.814870827Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:42.815798 containerd[1625]: time="2026-01-24T11:44:42.815749956Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 24 11:44:42.897429 containerd[1625]: time="2026-01-24T11:44:42.896805822Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:42.919446 containerd[1625]: time="2026-01-24T11:44:42.917502142Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 24 11:44:42.920343 containerd[1625]: time="2026-01-24T11:44:42.918599799Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:42.922364 kubelet[2999]: E0124 11:44:42.921702 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:44:42.922364 kubelet[2999]: E0124 11:44:42.921820 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:44:42.922364 kubelet[2999]: E0124 11:44:42.921936 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker start failed in pod whisker-65998ccc64-lw4d6_calico-system(c2801bc8-955a-42fc-b0a7-f5018299bf2f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:42.949490 containerd[1625]: time="2026-01-24T11:44:42.945909494Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 24 11:44:43.055886 containerd[1625]: time="2026-01-24T11:44:43.055686533Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:43.062390 containerd[1625]: time="2026-01-24T11:44:43.061977095Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 24 11:44:43.062835 containerd[1625]: time="2026-01-24T11:44:43.062660871Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:43.063786 kubelet[2999]: E0124 11:44:43.063525 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:44:43.063786 kubelet[2999]: E0124 11:44:43.063683 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:44:43.063991 kubelet[2999]: E0124 11:44:43.063802 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker-backend start failed in pod whisker-65998ccc64-lw4d6_calico-system(c2801bc8-955a-42fc-b0a7-f5018299bf2f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:43.063991 kubelet[2999]: E0124 11:44:43.063862 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:44:43.682000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.26:22-10.0.0.1:53856 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:43.683114 systemd[1]: Started sshd@15-10.0.0.26:22-10.0.0.1:53856.service - OpenSSH per-connection server daemon (10.0.0.1:53856). Jan 24 11:44:43.693413 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:44:43.693581 kernel: audit: type=1130 audit(1769255083.682:804): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.26:22-10.0.0.1:53856 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:43.949000 audit[5937]: USER_ACCT pid=5937 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:43.955085 sshd[5937]: Accepted publickey for core from 10.0.0.1 port 53856 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:43.960862 sshd-session[5937]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:43.979704 systemd-logind[1579]: New session 17 of user core. Jan 24 11:44:43.957000 audit[5937]: CRED_ACQ pid=5937 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.028136 kernel: audit: type=1101 audit(1769255083.949:805): pid=5937 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.028496 kernel: audit: type=1103 audit(1769255083.957:806): pid=5937 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.028549 kernel: audit: type=1006 audit(1769255083.957:807): pid=5937 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=17 res=1 Jan 24 11:44:43.957000 audit[5937]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdb802b560 a2=3 a3=0 items=0 ppid=1 pid=5937 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:44.081408 kernel: audit: type=1300 audit(1769255083.957:807): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdb802b560 a2=3 a3=0 items=0 ppid=1 pid=5937 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:44.081619 kernel: audit: type=1327 audit(1769255083.957:807): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:43.957000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:44.101703 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 24 11:44:44.113000 audit[5937]: USER_START pid=5937 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.120000 audit[5941]: CRED_ACQ pid=5941 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.166351 kernel: audit: type=1105 audit(1769255084.113:808): pid=5937 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.166473 kernel: audit: type=1103 audit(1769255084.120:809): pid=5941 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.538359 sshd[5941]: Connection closed by 10.0.0.1 port 53856 Jan 24 11:44:44.539679 sshd-session[5937]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:44.544000 audit[5937]: USER_END pid=5937 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.544000 audit[5937]: CRED_DISP pid=5937 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.614678 kernel: audit: type=1106 audit(1769255084.544:810): pid=5937 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.614822 kernel: audit: type=1104 audit(1769255084.544:811): pid=5937 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.626930 systemd[1]: sshd@15-10.0.0.26:22-10.0.0.1:53856.service: Deactivated successfully. Jan 24 11:44:44.626000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.26:22-10.0.0.1:53856 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:44.631095 systemd[1]: session-17.scope: Deactivated successfully. Jan 24 11:44:44.636135 systemd-logind[1579]: Session 17 logged out. Waiting for processes to exit. Jan 24 11:44:44.641000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.26:22-10.0.0.1:53868 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:44.642589 systemd[1]: Started sshd@16-10.0.0.26:22-10.0.0.1:53868.service - OpenSSH per-connection server daemon (10.0.0.1:53868). Jan 24 11:44:44.644905 systemd-logind[1579]: Removed session 17. Jan 24 11:44:44.813000 audit[5955]: USER_ACCT pid=5955 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.815579 sshd[5955]: Accepted publickey for core from 10.0.0.1 port 53868 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:44.822000 audit[5955]: CRED_ACQ pid=5955 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.822000 audit[5955]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcc9767800 a2=3 a3=0 items=0 ppid=1 pid=5955 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:44.822000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:44.827823 sshd-session[5955]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:44.853350 systemd-logind[1579]: New session 18 of user core. Jan 24 11:44:44.869825 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 24 11:44:44.885000 audit[5955]: USER_START pid=5955 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:44.892000 audit[5959]: CRED_ACQ pid=5959 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:45.329669 sshd[5959]: Connection closed by 10.0.0.1 port 53868 Jan 24 11:44:45.332545 sshd-session[5955]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:45.339000 audit[5955]: USER_END pid=5955 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:45.341000 audit[5955]: CRED_DISP pid=5955 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:45.349000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.26:22-10.0.0.1:53878 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:45.351000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.26:22-10.0.0.1:53868 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:45.350771 systemd[1]: Started sshd@17-10.0.0.26:22-10.0.0.1:53878.service - OpenSSH per-connection server daemon (10.0.0.1:53878). Jan 24 11:44:45.351772 systemd[1]: sshd@16-10.0.0.26:22-10.0.0.1:53868.service: Deactivated successfully. Jan 24 11:44:45.360954 systemd[1]: session-18.scope: Deactivated successfully. Jan 24 11:44:45.366631 systemd-logind[1579]: Session 18 logged out. Waiting for processes to exit. Jan 24 11:44:45.369767 systemd-logind[1579]: Removed session 18. Jan 24 11:44:45.523000 audit[5967]: USER_ACCT pid=5967 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:45.525368 sshd[5967]: Accepted publickey for core from 10.0.0.1 port 53878 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:45.527000 audit[5967]: CRED_ACQ pid=5967 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:45.528000 audit[5967]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd10747c90 a2=3 a3=0 items=0 ppid=1 pid=5967 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:45.528000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:45.537599 sshd-session[5967]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:45.555826 systemd-logind[1579]: New session 19 of user core. Jan 24 11:44:45.568996 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 24 11:44:45.577000 audit[5967]: USER_START pid=5967 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:45.583000 audit[5976]: CRED_ACQ pid=5976 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:45.941858 sshd[5976]: Connection closed by 10.0.0.1 port 53878 Jan 24 11:44:45.943524 sshd-session[5967]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:45.953000 audit[5967]: USER_END pid=5967 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:45.953000 audit[5967]: CRED_DISP pid=5967 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:45.964357 systemd[1]: sshd@17-10.0.0.26:22-10.0.0.1:53878.service: Deactivated successfully. Jan 24 11:44:45.964000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.26:22-10.0.0.1:53878 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:45.970974 systemd[1]: session-19.scope: Deactivated successfully. Jan 24 11:44:45.977521 systemd-logind[1579]: Session 19 logged out. Waiting for processes to exit. Jan 24 11:44:45.980917 systemd-logind[1579]: Removed session 19. Jan 24 11:44:49.674968 containerd[1625]: time="2026-01-24T11:44:49.674656650Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:44:49.762576 containerd[1625]: time="2026-01-24T11:44:49.762125046Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:49.774360 containerd[1625]: time="2026-01-24T11:44:49.774145901Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:44:49.774719 containerd[1625]: time="2026-01-24T11:44:49.774387423Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:49.775989 kubelet[2999]: E0124 11:44:49.775164 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:49.775989 kubelet[2999]: E0124 11:44:49.775889 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:44:49.776919 kubelet[2999]: E0124 11:44:49.776095 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:49.776919 kubelet[2999]: E0124 11:44:49.776153 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:44:50.969689 systemd[1]: Started sshd@18-10.0.0.26:22-10.0.0.1:53882.service - OpenSSH per-connection server daemon (10.0.0.1:53882). Jan 24 11:44:50.968000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.26:22-10.0.0.1:53882 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:51.008445 kernel: kauditd_printk_skb: 23 callbacks suppressed Jan 24 11:44:51.008611 kernel: audit: type=1130 audit(1769255090.968:831): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.26:22-10.0.0.1:53882 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:51.172000 audit[5997]: USER_ACCT pid=5997 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.179442 sshd[5997]: Accepted publickey for core from 10.0.0.1 port 53882 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:51.181120 sshd-session[5997]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:51.198295 systemd-logind[1579]: New session 20 of user core. Jan 24 11:44:51.211418 kernel: audit: type=1101 audit(1769255091.172:832): pid=5997 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.211539 kernel: audit: type=1103 audit(1769255091.176:833): pid=5997 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.176000 audit[5997]: CRED_ACQ pid=5997 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.241645 kernel: audit: type=1006 audit(1769255091.176:834): pid=5997 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=20 res=1 Jan 24 11:44:51.176000 audit[5997]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffde386cf20 a2=3 a3=0 items=0 ppid=1 pid=5997 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:51.283683 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 24 11:44:51.176000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:51.337934 kernel: audit: type=1300 audit(1769255091.176:834): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffde386cf20 a2=3 a3=0 items=0 ppid=1 pid=5997 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:51.338145 kernel: audit: type=1327 audit(1769255091.176:834): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:51.293000 audit[5997]: USER_START pid=5997 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.377093 kernel: audit: type=1105 audit(1769255091.293:835): pid=5997 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.308000 audit[6001]: CRED_ACQ pid=6001 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.418888 kernel: audit: type=1103 audit(1769255091.308:836): pid=6001 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.721397 sshd[6001]: Connection closed by 10.0.0.1 port 53882 Jan 24 11:44:51.721831 sshd-session[5997]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:51.726000 audit[5997]: USER_END pid=5997 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.736769 systemd[1]: sshd@18-10.0.0.26:22-10.0.0.1:53882.service: Deactivated successfully. Jan 24 11:44:51.740940 systemd[1]: session-20.scope: Deactivated successfully. Jan 24 11:44:51.743868 systemd-logind[1579]: Session 20 logged out. Waiting for processes to exit. Jan 24 11:44:51.748141 systemd-logind[1579]: Removed session 20. Jan 24 11:44:51.730000 audit[5997]: CRED_DISP pid=5997 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.767439 kernel: audit: type=1106 audit(1769255091.726:837): pid=5997 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.767485 kernel: audit: type=1104 audit(1769255091.730:838): pid=5997 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:51.736000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.26:22-10.0.0.1:53882 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:53.664816 kubelet[2999]: E0124 11:44:53.663665 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:53.679431 kubelet[2999]: E0124 11:44:53.679016 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:44:53.679431 kubelet[2999]: E0124 11:44:53.679494 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:44:53.679431 kubelet[2999]: E0124 11:44:53.679625 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:44:54.668355 kubelet[2999]: E0124 11:44:54.668111 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:44:55.673941 kubelet[2999]: E0124 11:44:55.672429 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:44:55.677176 containerd[1625]: time="2026-01-24T11:44:55.674451999Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 24 11:44:55.791414 containerd[1625]: time="2026-01-24T11:44:55.790705168Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:55.800987 containerd[1625]: time="2026-01-24T11:44:55.796952347Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 24 11:44:55.800987 containerd[1625]: time="2026-01-24T11:44:55.797313431Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:55.809581 kubelet[2999]: E0124 11:44:55.803688 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:44:55.809581 kubelet[2999]: E0124 11:44:55.803748 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:44:55.809581 kubelet[2999]: E0124 11:44:55.803844 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-csi start failed in pod csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:55.822777 containerd[1625]: time="2026-01-24T11:44:55.822737846Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 24 11:44:55.916734 containerd[1625]: time="2026-01-24T11:44:55.916539914Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:44:55.925509 containerd[1625]: time="2026-01-24T11:44:55.924126858Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 24 11:44:55.925509 containerd[1625]: time="2026-01-24T11:44:55.924462053Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 24 11:44:55.927745 kubelet[2999]: E0124 11:44:55.926923 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:44:55.927745 kubelet[2999]: E0124 11:44:55.927129 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:44:55.927745 kubelet[2999]: E0124 11:44:55.927390 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container csi-node-driver-registrar start failed in pod csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 24 11:44:55.927745 kubelet[2999]: E0124 11:44:55.927455 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:44:56.770842 systemd[1]: Started sshd@19-10.0.0.26:22-10.0.0.1:51586.service - OpenSSH per-connection server daemon (10.0.0.1:51586). Jan 24 11:44:56.769000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.26:22-10.0.0.1:51586 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:56.778446 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:44:56.778571 kernel: audit: type=1130 audit(1769255096.769:840): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.26:22-10.0.0.1:51586 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:56.964000 audit[6018]: USER_ACCT pid=6018 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:56.972567 sshd[6018]: Accepted publickey for core from 10.0.0.1 port 51586 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:44:56.976600 sshd-session[6018]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:44:56.995858 kernel: audit: type=1101 audit(1769255096.964:841): pid=6018 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:56.996378 kernel: audit: type=1103 audit(1769255096.968:842): pid=6018 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:56.968000 audit[6018]: CRED_ACQ pid=6018 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:57.024335 systemd-logind[1579]: New session 21 of user core. Jan 24 11:44:57.048642 kernel: audit: type=1006 audit(1769255096.968:843): pid=6018 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=21 res=1 Jan 24 11:44:57.048760 kernel: audit: type=1300 audit(1769255096.968:843): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffec37e8a30 a2=3 a3=0 items=0 ppid=1 pid=6018 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:56.968000 audit[6018]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffec37e8a30 a2=3 a3=0 items=0 ppid=1 pid=6018 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:44:57.077467 kernel: audit: type=1327 audit(1769255096.968:843): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:56.968000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:44:57.090973 systemd[1]: Started session-21.scope - Session 21 of User core. Jan 24 11:44:57.101000 audit[6018]: USER_START pid=6018 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:57.136407 kernel: audit: type=1105 audit(1769255097.101:844): pid=6018 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:57.136548 kernel: audit: type=1103 audit(1769255097.112:845): pid=6022 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:57.112000 audit[6022]: CRED_ACQ pid=6022 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:57.344590 sshd[6022]: Connection closed by 10.0.0.1 port 51586 Jan 24 11:44:57.346444 sshd-session[6018]: pam_unix(sshd:session): session closed for user core Jan 24 11:44:57.351000 audit[6018]: USER_END pid=6018 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:57.359382 systemd[1]: sshd@19-10.0.0.26:22-10.0.0.1:51586.service: Deactivated successfully. Jan 24 11:44:57.364418 systemd[1]: session-21.scope: Deactivated successfully. Jan 24 11:44:57.371687 systemd-logind[1579]: Session 21 logged out. Waiting for processes to exit. Jan 24 11:44:57.376103 systemd-logind[1579]: Removed session 21. Jan 24 11:44:57.351000 audit[6018]: CRED_DISP pid=6018 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:57.413697 kernel: audit: type=1106 audit(1769255097.351:846): pid=6018 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:57.413904 kernel: audit: type=1104 audit(1769255097.351:847): pid=6018 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:44:57.356000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.26:22-10.0.0.1:51586 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:44:58.667438 kubelet[2999]: E0124 11:44:58.664473 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:45:02.395000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.26:22-10.0.0.1:58768 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:02.396789 systemd[1]: Started sshd@20-10.0.0.26:22-10.0.0.1:58768.service - OpenSSH per-connection server daemon (10.0.0.1:58768). Jan 24 11:45:02.413343 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:45:02.413450 kernel: audit: type=1130 audit(1769255102.395:849): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.26:22-10.0.0.1:58768 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:02.620000 audit[6067]: USER_ACCT pid=6067 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:02.629817 sshd-session[6067]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:02.639766 sshd[6067]: Accepted publickey for core from 10.0.0.1 port 58768 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:02.655692 kernel: audit: type=1101 audit(1769255102.620:850): pid=6067 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:02.624000 audit[6067]: CRED_ACQ pid=6067 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:02.670155 kubelet[2999]: E0124 11:45:02.669006 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:45:02.681799 kubelet[2999]: E0124 11:45:02.681623 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:45:02.689412 kernel: audit: type=1103 audit(1769255102.624:851): pid=6067 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:02.713574 kernel: audit: type=1006 audit(1769255102.625:852): pid=6067 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=22 res=1 Jan 24 11:45:02.713417 systemd-logind[1579]: New session 22 of user core. Jan 24 11:45:02.625000 audit[6067]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff778ab530 a2=3 a3=0 items=0 ppid=1 pid=6067 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:02.750293 kernel: audit: type=1300 audit(1769255102.625:852): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff778ab530 a2=3 a3=0 items=0 ppid=1 pid=6067 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:02.625000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:02.763426 kernel: audit: type=1327 audit(1769255102.625:852): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:02.764760 systemd[1]: Started session-22.scope - Session 22 of User core. Jan 24 11:45:02.779000 audit[6067]: USER_START pid=6067 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:02.829353 kernel: audit: type=1105 audit(1769255102.779:853): pid=6067 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:02.829489 kernel: audit: type=1103 audit(1769255102.791:854): pid=6073 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:02.791000 audit[6073]: CRED_ACQ pid=6073 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:03.214546 sshd[6073]: Connection closed by 10.0.0.1 port 58768 Jan 24 11:45:03.215583 sshd-session[6067]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:03.227000 audit[6067]: USER_END pid=6067 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:03.278631 kernel: audit: type=1106 audit(1769255103.227:855): pid=6067 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:03.278443 systemd[1]: sshd@20-10.0.0.26:22-10.0.0.1:58768.service: Deactivated successfully. Jan 24 11:45:03.227000 audit[6067]: CRED_DISP pid=6067 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:03.284777 systemd[1]: session-22.scope: Deactivated successfully. Jan 24 11:45:03.291746 systemd-logind[1579]: Session 22 logged out. Waiting for processes to exit. Jan 24 11:45:03.300951 systemd-logind[1579]: Removed session 22. Jan 24 11:45:03.328786 kernel: audit: type=1104 audit(1769255103.227:856): pid=6067 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:03.277000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.26:22-10.0.0.1:58768 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:04.662485 kubelet[2999]: E0124 11:45:04.661924 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:45:04.666527 kubelet[2999]: E0124 11:45:04.665744 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:45:05.669709 kubelet[2999]: E0124 11:45:05.669570 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:45:06.671610 kubelet[2999]: E0124 11:45:06.670954 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:45:07.667497 kubelet[2999]: E0124 11:45:07.666717 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:45:08.255970 systemd[1]: Started sshd@21-10.0.0.26:22-10.0.0.1:58770.service - OpenSSH per-connection server daemon (10.0.0.1:58770). Jan 24 11:45:08.254000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.26:22-10.0.0.1:58770 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:08.264441 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:45:08.264517 kernel: audit: type=1130 audit(1769255108.254:858): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.26:22-10.0.0.1:58770 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:08.535029 sshd[6091]: Accepted publickey for core from 10.0.0.1 port 58770 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:08.533000 audit[6091]: USER_ACCT pid=6091 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:08.546613 sshd-session[6091]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:08.563603 systemd-logind[1579]: New session 23 of user core. Jan 24 11:45:08.542000 audit[6091]: CRED_ACQ pid=6091 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:08.617920 kernel: audit: type=1101 audit(1769255108.533:859): pid=6091 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:08.618372 kernel: audit: type=1103 audit(1769255108.542:860): pid=6091 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:08.618458 kernel: audit: type=1006 audit(1769255108.542:861): pid=6091 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=23 res=1 Jan 24 11:45:08.542000 audit[6091]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc22a24c30 a2=3 a3=0 items=0 ppid=1 pid=6091 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:08.678438 kernel: audit: type=1300 audit(1769255108.542:861): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc22a24c30 a2=3 a3=0 items=0 ppid=1 pid=6091 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:08.678548 kernel: audit: type=1327 audit(1769255108.542:861): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:08.542000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:08.672800 systemd[1]: Started session-23.scope - Session 23 of User core. Jan 24 11:45:08.691557 kernel: audit: type=1105 audit(1769255108.680:862): pid=6091 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:08.680000 audit[6091]: USER_START pid=6091 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:08.684000 audit[6095]: CRED_ACQ pid=6095 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:08.767633 kernel: audit: type=1103 audit(1769255108.684:863): pid=6095 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:09.060875 sshd[6095]: Connection closed by 10.0.0.1 port 58770 Jan 24 11:45:09.062743 sshd-session[6091]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:09.068000 audit[6091]: USER_END pid=6091 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:09.078437 systemd[1]: sshd@21-10.0.0.26:22-10.0.0.1:58770.service: Deactivated successfully. Jan 24 11:45:09.083662 systemd[1]: session-23.scope: Deactivated successfully. Jan 24 11:45:09.087166 systemd-logind[1579]: Session 23 logged out. Waiting for processes to exit. Jan 24 11:45:09.093516 systemd-logind[1579]: Removed session 23. Jan 24 11:45:09.069000 audit[6091]: CRED_DISP pid=6091 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:09.130569 kernel: audit: type=1106 audit(1769255109.068:864): pid=6091 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:09.130705 kernel: audit: type=1104 audit(1769255109.069:865): pid=6091 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:09.079000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.26:22-10.0.0.1:58770 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:11.666720 kubelet[2999]: E0124 11:45:11.666517 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:45:14.119548 systemd[1]: Started sshd@22-10.0.0.26:22-10.0.0.1:41998.service - OpenSSH per-connection server daemon (10.0.0.1:41998). Jan 24 11:45:14.120000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.26:22-10.0.0.1:41998 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:14.129523 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:45:14.129611 kernel: audit: type=1130 audit(1769255114.120:867): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.26:22-10.0.0.1:41998 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:14.322000 audit[6108]: USER_ACCT pid=6108 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.324764 sshd[6108]: Accepted publickey for core from 10.0.0.1 port 41998 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:14.332016 sshd-session[6108]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:14.364490 kernel: audit: type=1101 audit(1769255114.322:868): pid=6108 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.326000 audit[6108]: CRED_ACQ pid=6108 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.364928 systemd-logind[1579]: New session 24 of user core. Jan 24 11:45:14.438479 kernel: audit: type=1103 audit(1769255114.326:869): pid=6108 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.439602 kernel: audit: type=1006 audit(1769255114.327:870): pid=6108 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=24 res=1 Jan 24 11:45:14.439663 kernel: audit: type=1300 audit(1769255114.327:870): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd41d629b0 a2=3 a3=0 items=0 ppid=1 pid=6108 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:14.327000 audit[6108]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd41d629b0 a2=3 a3=0 items=0 ppid=1 pid=6108 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:14.327000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:14.491455 kernel: audit: type=1327 audit(1769255114.327:870): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:14.497484 systemd[1]: Started session-24.scope - Session 24 of User core. Jan 24 11:45:14.521000 audit[6108]: USER_START pid=6108 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.569914 kernel: audit: type=1105 audit(1769255114.521:871): pid=6108 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.570038 kernel: audit: type=1103 audit(1769255114.529:872): pid=6112 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.529000 audit[6112]: CRED_ACQ pid=6112 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.671811 kubelet[2999]: E0124 11:45:14.671707 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:45:14.966633 sshd[6112]: Connection closed by 10.0.0.1 port 41998 Jan 24 11:45:14.966480 sshd-session[6108]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:14.969000 audit[6108]: USER_END pid=6108 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.979367 systemd[1]: sshd@22-10.0.0.26:22-10.0.0.1:41998.service: Deactivated successfully. Jan 24 11:45:14.986051 systemd[1]: session-24.scope: Deactivated successfully. Jan 24 11:45:14.992769 systemd-logind[1579]: Session 24 logged out. Waiting for processes to exit. Jan 24 11:45:14.996557 systemd-logind[1579]: Removed session 24. Jan 24 11:45:14.971000 audit[6108]: CRED_DISP pid=6108 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:15.046407 kernel: audit: type=1106 audit(1769255114.969:873): pid=6108 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:15.046573 kernel: audit: type=1104 audit(1769255114.971:874): pid=6108 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:14.976000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.26:22-10.0.0.1:41998 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:15.665077 kubelet[2999]: E0124 11:45:15.663571 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:45:17.663953 kubelet[2999]: E0124 11:45:17.663814 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:45:17.667014 kubelet[2999]: E0124 11:45:17.666842 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:45:19.668414 containerd[1625]: time="2026-01-24T11:45:19.665789759Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 24 11:45:19.775301 containerd[1625]: time="2026-01-24T11:45:19.775017736Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:45:19.779639 containerd[1625]: time="2026-01-24T11:45:19.779483153Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 24 11:45:19.779757 containerd[1625]: time="2026-01-24T11:45:19.779659211Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 24 11:45:19.780703 kubelet[2999]: E0124 11:45:19.780444 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:45:19.780703 kubelet[2999]: E0124 11:45:19.780568 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:45:19.780703 kubelet[2999]: E0124 11:45:19.780654 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-kube-controllers start failed in pod calico-kube-controllers-7dbccbb54b-mvbw5_calico-system(c101de33-34a1-459f-bcfc-2378435cf35e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 24 11:45:19.780703 kubelet[2999]: E0124 11:45:19.780689 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:45:19.994000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.26:22-10.0.0.1:42012 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:19.994931 systemd[1]: Started sshd@23-10.0.0.26:22-10.0.0.1:42012.service - OpenSSH per-connection server daemon (10.0.0.1:42012). Jan 24 11:45:20.009532 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:45:20.009599 kernel: audit: type=1130 audit(1769255119.994:876): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.26:22-10.0.0.1:42012 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:20.172000 audit[6126]: USER_ACCT pid=6126 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.195968 sshd[6126]: Accepted publickey for core from 10.0.0.1 port 42012 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:20.206456 sshd-session[6126]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:20.212696 kernel: audit: type=1101 audit(1769255120.172:877): pid=6126 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.199000 audit[6126]: CRED_ACQ pid=6126 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.234991 systemd-logind[1579]: New session 25 of user core. Jan 24 11:45:20.252772 kernel: audit: type=1103 audit(1769255120.199:878): pid=6126 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.284741 kernel: audit: type=1006 audit(1769255120.199:879): pid=6126 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=25 res=1 Jan 24 11:45:20.284877 kernel: audit: type=1300 audit(1769255120.199:879): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffeb03b72b0 a2=3 a3=0 items=0 ppid=1 pid=6126 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:20.199000 audit[6126]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffeb03b72b0 a2=3 a3=0 items=0 ppid=1 pid=6126 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:20.290938 systemd[1]: Started session-25.scope - Session 25 of User core. Jan 24 11:45:20.199000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:20.346525 kernel: audit: type=1327 audit(1769255120.199:879): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:20.312000 audit[6126]: USER_START pid=6126 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.399641 kernel: audit: type=1105 audit(1769255120.312:880): pid=6126 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.322000 audit[6130]: CRED_ACQ pid=6130 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.433837 kernel: audit: type=1103 audit(1769255120.322:881): pid=6130 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.657751 sshd[6130]: Connection closed by 10.0.0.1 port 42012 Jan 24 11:45:20.660561 sshd-session[6126]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:20.664000 audit[6126]: USER_END pid=6126 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.664000 audit[6126]: CRED_DISP pid=6126 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.711494 kernel: audit: type=1106 audit(1769255120.664:882): pid=6126 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.711585 kernel: audit: type=1104 audit(1769255120.664:883): pid=6126 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.721414 systemd[1]: sshd@23-10.0.0.26:22-10.0.0.1:42012.service: Deactivated successfully. Jan 24 11:45:20.731521 systemd[1]: session-25.scope: Deactivated successfully. Jan 24 11:45:20.737825 systemd-logind[1579]: Session 25 logged out. Waiting for processes to exit. Jan 24 11:45:20.745525 systemd[1]: Started sshd@24-10.0.0.26:22-10.0.0.1:42014.service - OpenSSH per-connection server daemon (10.0.0.1:42014). Jan 24 11:45:20.722000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.26:22-10.0.0.1:42012 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:20.744000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.26:22-10.0.0.1:42014 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:20.754616 systemd-logind[1579]: Removed session 25. Jan 24 11:45:20.896000 audit[6144]: USER_ACCT pid=6144 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.898792 sshd[6144]: Accepted publickey for core from 10.0.0.1 port 42014 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:20.904000 audit[6144]: CRED_ACQ pid=6144 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.904000 audit[6144]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffec15b4af0 a2=3 a3=0 items=0 ppid=1 pid=6144 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=26 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:20.904000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:20.908566 sshd-session[6144]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:20.931712 systemd-logind[1579]: New session 26 of user core. Jan 24 11:45:20.948442 systemd[1]: Started session-26.scope - Session 26 of User core. Jan 24 11:45:20.962000 audit[6144]: USER_START pid=6144 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:20.967000 audit[6149]: CRED_ACQ pid=6149 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:22.460950 sshd[6149]: Connection closed by 10.0.0.1 port 42014 Jan 24 11:45:22.465451 sshd-session[6144]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:22.533000 audit[6144]: USER_END pid=6144 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:22.533000 audit[6144]: CRED_DISP pid=6144 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:22.544588 systemd[1]: Started sshd@25-10.0.0.26:22-10.0.0.1:48414.service - OpenSSH per-connection server daemon (10.0.0.1:48414). Jan 24 11:45:22.543000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.26:22-10.0.0.1:48414 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:22.552728 systemd[1]: sshd@24-10.0.0.26:22-10.0.0.1:42014.service: Deactivated successfully. Jan 24 11:45:22.552000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.26:22-10.0.0.1:42014 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:22.566046 systemd[1]: session-26.scope: Deactivated successfully. Jan 24 11:45:22.586544 systemd-logind[1579]: Session 26 logged out. Waiting for processes to exit. Jan 24 11:45:22.595722 systemd-logind[1579]: Removed session 26. Jan 24 11:45:22.669742 kubelet[2999]: E0124 11:45:22.669515 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:45:22.927000 audit[6157]: USER_ACCT pid=6157 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:22.930528 sshd[6157]: Accepted publickey for core from 10.0.0.1 port 48414 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:22.935000 audit[6157]: CRED_ACQ pid=6157 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:22.937000 audit[6157]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffda63a69f0 a2=3 a3=0 items=0 ppid=1 pid=6157 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=27 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:22.937000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:22.942692 sshd-session[6157]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:22.979894 systemd-logind[1579]: New session 27 of user core. Jan 24 11:45:23.019808 systemd[1]: Started session-27.scope - Session 27 of User core. Jan 24 11:45:23.043000 audit[6157]: USER_START pid=6157 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:23.054000 audit[6164]: CRED_ACQ pid=6164 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:24.680444 kubelet[2999]: E0124 11:45:24.679721 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:45:24.903000 audit[6181]: NETFILTER_CFG table=filter:141 family=2 entries=26 op=nft_register_rule pid=6181 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:45:24.903000 audit[6181]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffdc4314850 a2=0 a3=7ffdc431483c items=0 ppid=3157 pid=6181 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:24.903000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:45:24.923000 audit[6181]: NETFILTER_CFG table=nat:142 family=2 entries=20 op=nft_register_rule pid=6181 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:45:24.923000 audit[6181]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffdc4314850 a2=0 a3=0 items=0 ppid=3157 pid=6181 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:24.923000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:45:24.975471 sshd[6164]: Connection closed by 10.0.0.1 port 48414 Jan 24 11:45:24.976069 sshd-session[6157]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:24.980000 audit[6157]: USER_END pid=6157 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:24.980000 audit[6157]: CRED_DISP pid=6157 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:25.024590 kernel: kauditd_printk_skb: 28 callbacks suppressed Jan 24 11:45:25.024732 kernel: audit: type=1130 audit(1769255124.997:904): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.26:22-10.0.0.1:48422 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:24.997000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.26:22-10.0.0.1:48422 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:24.998894 systemd[1]: Started sshd@26-10.0.0.26:22-10.0.0.1:48422.service - OpenSSH per-connection server daemon (10.0.0.1:48422). Jan 24 11:45:25.000652 systemd[1]: sshd@25-10.0.0.26:22-10.0.0.1:48414.service: Deactivated successfully. Jan 24 11:45:25.045778 systemd[1]: session-27.scope: Deactivated successfully. Jan 24 11:45:25.047626 systemd[1]: session-27.scope: Consumed 1.221s CPU time, 39.8M memory peak. Jan 24 11:45:25.011000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.26:22-10.0.0.1:48414 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:25.053592 systemd-logind[1579]: Session 27 logged out. Waiting for processes to exit. Jan 24 11:45:25.062427 systemd-logind[1579]: Removed session 27. Jan 24 11:45:25.086928 kernel: audit: type=1131 audit(1769255125.011:905): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.26:22-10.0.0.1:48414 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:25.365000 audit[6183]: USER_ACCT pid=6183 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:25.368374 sshd[6183]: Accepted publickey for core from 10.0.0.1 port 48422 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:25.372329 sshd-session[6183]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:25.397844 systemd-logind[1579]: New session 28 of user core. Jan 24 11:45:25.369000 audit[6183]: CRED_ACQ pid=6183 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:25.438815 kernel: audit: type=1101 audit(1769255125.365:906): pid=6183 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:25.438942 kernel: audit: type=1103 audit(1769255125.369:907): pid=6183 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:25.462682 kernel: audit: type=1006 audit(1769255125.369:908): pid=6183 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=28 res=1 Jan 24 11:45:25.463888 systemd[1]: Started session-28.scope - Session 28 of User core. Jan 24 11:45:25.369000 audit[6183]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc20c1e030 a2=3 a3=0 items=0 ppid=1 pid=6183 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=28 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:25.513018 kernel: audit: type=1300 audit(1769255125.369:908): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc20c1e030 a2=3 a3=0 items=0 ppid=1 pid=6183 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=28 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:25.513557 kernel: audit: type=1327 audit(1769255125.369:908): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:25.369000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:25.479000 audit[6183]: USER_START pid=6183 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:25.569071 kernel: audit: type=1105 audit(1769255125.479:909): pid=6183 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:25.485000 audit[6191]: CRED_ACQ pid=6191 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:25.611771 kernel: audit: type=1103 audit(1769255125.485:910): pid=6191 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:26.018000 audit[6206]: NETFILTER_CFG table=filter:143 family=2 entries=38 op=nft_register_rule pid=6206 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:45:26.054727 kernel: audit: type=1325 audit(1769255126.018:911): table=filter:143 family=2 entries=38 op=nft_register_rule pid=6206 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:45:26.018000 audit[6206]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffc8d0fa0f0 a2=0 a3=7ffc8d0fa0dc items=0 ppid=3157 pid=6206 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:26.018000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:45:26.065000 audit[6206]: NETFILTER_CFG table=nat:144 family=2 entries=20 op=nft_register_rule pid=6206 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:45:26.065000 audit[6206]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffc8d0fa0f0 a2=0 a3=0 items=0 ppid=3157 pid=6206 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:26.065000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:45:26.430943 sshd[6191]: Connection closed by 10.0.0.1 port 48422 Jan 24 11:45:26.430095 sshd-session[6183]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:26.442000 audit[6183]: USER_END pid=6183 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:26.442000 audit[6183]: CRED_DISP pid=6183 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:26.464928 systemd[1]: Started sshd@27-10.0.0.26:22-10.0.0.1:48432.service - OpenSSH per-connection server daemon (10.0.0.1:48432). Jan 24 11:45:26.465000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.26:22-10.0.0.1:48432 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:26.468484 systemd[1]: sshd@26-10.0.0.26:22-10.0.0.1:48422.service: Deactivated successfully. Jan 24 11:45:26.473000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.26:22-10.0.0.1:48422 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:26.483743 systemd[1]: session-28.scope: Deactivated successfully. Jan 24 11:45:26.490802 systemd-logind[1579]: Session 28 logged out. Waiting for processes to exit. Jan 24 11:45:26.495642 systemd-logind[1579]: Removed session 28. Jan 24 11:45:26.671000 audit[6208]: USER_ACCT pid=6208 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:26.672011 sshd[6208]: Accepted publickey for core from 10.0.0.1 port 48432 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:26.675000 audit[6208]: CRED_ACQ pid=6208 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:26.675000 audit[6208]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe496a1b90 a2=3 a3=0 items=0 ppid=1 pid=6208 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=29 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:26.675000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:26.680557 sshd-session[6208]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:26.718056 systemd-logind[1579]: New session 29 of user core. Jan 24 11:45:26.734502 systemd[1]: Started session-29.scope - Session 29 of User core. Jan 24 11:45:26.748000 audit[6208]: USER_START pid=6208 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:26.754000 audit[6215]: CRED_ACQ pid=6215 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:27.140391 sshd[6215]: Connection closed by 10.0.0.1 port 48432 Jan 24 11:45:27.141342 sshd-session[6208]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:27.155000 audit[6208]: USER_END pid=6208 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:27.155000 audit[6208]: CRED_DISP pid=6208 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:27.163969 systemd[1]: sshd@27-10.0.0.26:22-10.0.0.1:48432.service: Deactivated successfully. Jan 24 11:45:27.165000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.26:22-10.0.0.1:48432 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:27.171925 systemd[1]: session-29.scope: Deactivated successfully. Jan 24 11:45:27.175439 systemd-logind[1579]: Session 29 logged out. Waiting for processes to exit. Jan 24 11:45:27.180481 systemd-logind[1579]: Removed session 29. Jan 24 11:45:29.670624 containerd[1625]: time="2026-01-24T11:45:29.670459748Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 24 11:45:29.675915 kubelet[2999]: E0124 11:45:29.675855 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:45:29.775019 containerd[1625]: time="2026-01-24T11:45:29.771989523Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:45:29.778778 containerd[1625]: time="2026-01-24T11:45:29.778736586Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 24 11:45:29.778980 containerd[1625]: time="2026-01-24T11:45:29.778961434Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 24 11:45:29.779724 kubelet[2999]: E0124 11:45:29.779679 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:45:29.780454 kubelet[2999]: E0124 11:45:29.779835 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:45:29.780454 kubelet[2999]: E0124 11:45:29.780071 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker start failed in pod whisker-65998ccc64-lw4d6_calico-system(c2801bc8-955a-42fc-b0a7-f5018299bf2f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 24 11:45:29.785531 containerd[1625]: time="2026-01-24T11:45:29.785474590Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:45:29.878472 containerd[1625]: time="2026-01-24T11:45:29.873754328Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:45:29.881496 containerd[1625]: time="2026-01-24T11:45:29.880778281Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:45:29.881496 containerd[1625]: time="2026-01-24T11:45:29.880879800Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:45:29.885738 kubelet[2999]: E0124 11:45:29.883470 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:45:29.885738 kubelet[2999]: E0124 11:45:29.883610 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:45:29.885738 kubelet[2999]: E0124 11:45:29.883864 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-f58644cbc-hhctc_calico-apiserver(ec314a61-36d7-4af0-b9d6-11f6bfe86b8c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:45:29.885911 containerd[1625]: time="2026-01-24T11:45:29.884657800Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 24 11:45:29.886934 kubelet[2999]: E0124 11:45:29.886474 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:45:29.974745 containerd[1625]: time="2026-01-24T11:45:29.974687988Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:45:29.979967 containerd[1625]: time="2026-01-24T11:45:29.979918439Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 24 11:45:29.980901 containerd[1625]: time="2026-01-24T11:45:29.980372084Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 24 11:45:29.983634 kubelet[2999]: E0124 11:45:29.982785 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:45:29.983634 kubelet[2999]: E0124 11:45:29.982831 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:45:29.983634 kubelet[2999]: E0124 11:45:29.982901 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker-backend start failed in pod whisker-65998ccc64-lw4d6_calico-system(c2801bc8-955a-42fc-b0a7-f5018299bf2f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 24 11:45:29.983634 kubelet[2999]: E0124 11:45:29.982937 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:45:32.166949 systemd[1]: Started sshd@28-10.0.0.26:22-10.0.0.1:48434.service - OpenSSH per-connection server daemon (10.0.0.1:48434). Jan 24 11:45:32.169000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.26:22-10.0.0.1:48434 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:32.176878 kernel: kauditd_printk_skb: 19 callbacks suppressed Jan 24 11:45:32.176939 kernel: audit: type=1130 audit(1769255132.169:925): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.26:22-10.0.0.1:48434 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:32.339794 sshd[6254]: Accepted publickey for core from 10.0.0.1 port 48434 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:32.337000 audit[6254]: USER_ACCT pid=6254 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.345904 sshd-session[6254]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:32.369492 systemd-logind[1579]: New session 30 of user core. Jan 24 11:45:32.343000 audit[6254]: CRED_ACQ pid=6254 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.413430 kernel: audit: type=1101 audit(1769255132.337:926): pid=6254 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.413549 kernel: audit: type=1103 audit(1769255132.343:927): pid=6254 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.413627 kernel: audit: type=1006 audit(1769255132.343:928): pid=6254 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=30 res=1 Jan 24 11:45:32.343000 audit[6254]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd8cddb690 a2=3 a3=0 items=0 ppid=1 pid=6254 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=30 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:32.473699 kernel: audit: type=1300 audit(1769255132.343:928): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd8cddb690 a2=3 a3=0 items=0 ppid=1 pid=6254 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=30 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:32.473847 kernel: audit: type=1327 audit(1769255132.343:928): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:32.343000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:32.489362 systemd[1]: Started session-30.scope - Session 30 of User core. Jan 24 11:45:32.501000 audit[6254]: USER_START pid=6254 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.544528 kernel: audit: type=1105 audit(1769255132.501:929): pid=6254 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.544673 kernel: audit: type=1103 audit(1769255132.508:930): pid=6260 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.508000 audit[6260]: CRED_ACQ pid=6260 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.797737 sshd[6260]: Connection closed by 10.0.0.1 port 48434 Jan 24 11:45:32.819984 sshd-session[6254]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:32.825000 audit[6254]: USER_END pid=6254 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.834985 systemd[1]: sshd@28-10.0.0.26:22-10.0.0.1:48434.service: Deactivated successfully. Jan 24 11:45:32.843595 systemd[1]: session-30.scope: Deactivated successfully. Jan 24 11:45:32.848417 systemd-logind[1579]: Session 30 logged out. Waiting for processes to exit. Jan 24 11:45:32.851806 systemd-logind[1579]: Removed session 30. Jan 24 11:45:32.875403 kernel: audit: type=1106 audit(1769255132.825:931): pid=6254 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.875709 kernel: audit: type=1104 audit(1769255132.826:932): pid=6254 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.826000 audit[6254]: CRED_DISP pid=6254 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:32.835000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.26:22-10.0.0.1:48434 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:33.673612 kubelet[2999]: E0124 11:45:33.673004 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e" Jan 24 11:45:34.020000 audit[6274]: NETFILTER_CFG table=filter:145 family=2 entries=26 op=nft_register_rule pid=6274 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:45:34.020000 audit[6274]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffd9366dcc0 a2=0 a3=7ffd9366dcac items=0 ppid=3157 pid=6274 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:34.020000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:45:34.050000 audit[6274]: NETFILTER_CFG table=nat:146 family=2 entries=104 op=nft_register_chain pid=6274 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:45:34.050000 audit[6274]: SYSCALL arch=c000003e syscall=46 success=yes exit=48684 a0=3 a1=7ffd9366dcc0 a2=0 a3=7ffd9366dcac items=0 ppid=3157 pid=6274 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:34.050000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:45:37.674619 containerd[1625]: time="2026-01-24T11:45:37.672066780Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 24 11:45:37.750943 containerd[1625]: time="2026-01-24T11:45:37.750581127Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:45:37.755681 containerd[1625]: time="2026-01-24T11:45:37.754910446Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 24 11:45:37.755681 containerd[1625]: time="2026-01-24T11:45:37.755013888Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 24 11:45:37.757585 kubelet[2999]: E0124 11:45:37.757440 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:45:37.758001 kubelet[2999]: E0124 11:45:37.757591 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:45:37.758001 kubelet[2999]: E0124 11:45:37.757699 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container goldmane start failed in pod goldmane-7c778bb748-4gb25_calico-system(9f91eb4d-a73b-4530-82b4-682e0b2b659a): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 24 11:45:37.759173 kubelet[2999]: E0124 11:45:37.758931 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-4gb25" podUID="9f91eb4d-a73b-4530-82b4-682e0b2b659a" Jan 24 11:45:37.829876 systemd[1]: Started sshd@29-10.0.0.26:22-10.0.0.1:56704.service - OpenSSH per-connection server daemon (10.0.0.1:56704). Jan 24 11:45:37.830000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.26:22-10.0.0.1:56704 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:37.846589 kernel: kauditd_printk_skb: 7 callbacks suppressed Jan 24 11:45:37.846731 kernel: audit: type=1130 audit(1769255137.830:936): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.26:22-10.0.0.1:56704 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:38.121000 audit[6285]: USER_ACCT pid=6285 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.130522 sshd-session[6285]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:38.134381 sshd[6285]: Accepted publickey for core from 10.0.0.1 port 56704 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:38.150972 systemd-logind[1579]: New session 31 of user core. Jan 24 11:45:38.168492 kernel: audit: type=1101 audit(1769255138.121:937): pid=6285 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.126000 audit[6285]: CRED_ACQ pid=6285 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.234943 kernel: audit: type=1103 audit(1769255138.126:938): pid=6285 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.235359 kernel: audit: type=1006 audit(1769255138.126:939): pid=6285 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=31 res=1 Jan 24 11:45:38.126000 audit[6285]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd20533830 a2=3 a3=0 items=0 ppid=1 pid=6285 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=31 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:38.241475 systemd[1]: Started session-31.scope - Session 31 of User core. Jan 24 11:45:38.281832 kernel: audit: type=1300 audit(1769255138.126:939): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd20533830 a2=3 a3=0 items=0 ppid=1 pid=6285 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=31 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:38.283625 kernel: audit: type=1327 audit(1769255138.126:939): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:38.126000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:38.305587 kernel: audit: type=1105 audit(1769255138.249:940): pid=6285 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.249000 audit[6285]: USER_START pid=6285 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.254000 audit[6289]: CRED_ACQ pid=6289 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.392463 kernel: audit: type=1103 audit(1769255138.254:941): pid=6289 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.640606 sshd[6289]: Connection closed by 10.0.0.1 port 56704 Jan 24 11:45:38.641559 sshd-session[6285]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:38.645000 audit[6285]: USER_END pid=6285 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.670049 systemd[1]: sshd@29-10.0.0.26:22-10.0.0.1:56704.service: Deactivated successfully. Jan 24 11:45:38.682433 systemd[1]: session-31.scope: Deactivated successfully. Jan 24 11:45:38.697861 kernel: audit: type=1106 audit(1769255138.645:942): pid=6285 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.700738 systemd-logind[1579]: Session 31 logged out. Waiting for processes to exit. Jan 24 11:45:38.704967 systemd-logind[1579]: Removed session 31. Jan 24 11:45:38.645000 audit[6285]: CRED_DISP pid=6285 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:38.669000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.26:22-10.0.0.1:56704 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:38.750591 kernel: audit: type=1104 audit(1769255138.645:943): pid=6285 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:39.668410 containerd[1625]: time="2026-01-24T11:45:39.667835420Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 24 11:45:39.762623 containerd[1625]: time="2026-01-24T11:45:39.761640031Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:45:39.772410 containerd[1625]: time="2026-01-24T11:45:39.772359284Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 24 11:45:39.772604 containerd[1625]: time="2026-01-24T11:45:39.772498511Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 24 11:45:39.777422 kubelet[2999]: E0124 11:45:39.776802 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:45:39.778335 kubelet[2999]: E0124 11:45:39.777456 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:45:39.778335 kubelet[2999]: E0124 11:45:39.777565 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-csi start failed in pod csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 24 11:45:39.792428 containerd[1625]: time="2026-01-24T11:45:39.790962454Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 24 11:45:39.898755 containerd[1625]: time="2026-01-24T11:45:39.897567740Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:45:39.906566 containerd[1625]: time="2026-01-24T11:45:39.906412427Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 24 11:45:39.906566 containerd[1625]: time="2026-01-24T11:45:39.906530687Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 24 11:45:39.911428 kubelet[2999]: E0124 11:45:39.909639 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:45:39.911428 kubelet[2999]: E0124 11:45:39.909710 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:45:39.911428 kubelet[2999]: E0124 11:45:39.909807 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container csi-node-driver-registrar start failed in pod csi-node-driver-j5qxv_calico-system(9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 24 11:45:39.911428 kubelet[2999]: E0124 11:45:39.910672 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-j5qxv" podUID="9501b5f5-3a29-45b9-8cd6-9e7a8c6911fd" Jan 24 11:45:40.703485 kubelet[2999]: E0124 11:45:40.701839 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:45:41.668490 kubelet[2999]: E0124 11:45:41.667985 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-hhctc" podUID="ec314a61-36d7-4af0-b9d6-11f6bfe86b8c" Jan 24 11:45:42.667864 containerd[1625]: time="2026-01-24T11:45:42.667713012Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:45:42.678834 kubelet[2999]: E0124 11:45:42.678570 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-65998ccc64-lw4d6" podUID="c2801bc8-955a-42fc-b0a7-f5018299bf2f" Jan 24 11:45:42.782769 containerd[1625]: time="2026-01-24T11:45:42.781812162Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:45:42.788561 containerd[1625]: time="2026-01-24T11:45:42.788487335Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:45:42.788639 containerd[1625]: time="2026-01-24T11:45:42.788608049Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:45:42.790386 kubelet[2999]: E0124 11:45:42.790343 2999 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:45:42.790533 kubelet[2999]: E0124 11:45:42.790513 2999 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:45:42.790659 kubelet[2999]: E0124 11:45:42.790642 2999 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-f58644cbc-dvnfm_calico-apiserver(eda31195-c93b-4ab6-93d0-16b77a14ef46): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:45:42.790737 kubelet[2999]: E0124 11:45:42.790716 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f58644cbc-dvnfm" podUID="eda31195-c93b-4ab6-93d0-16b77a14ef46" Jan 24 11:45:43.690848 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:45:43.690983 kernel: audit: type=1130 audit(1769255143.669:945): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@30-10.0.0.26:22-10.0.0.1:40578 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:43.669000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@30-10.0.0.26:22-10.0.0.1:40578 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:43.670741 systemd[1]: Started sshd@30-10.0.0.26:22-10.0.0.1:40578.service - OpenSSH per-connection server daemon (10.0.0.1:40578). Jan 24 11:45:43.857000 audit[6316]: USER_ACCT pid=6316 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:43.857823 sshd[6316]: Accepted publickey for core from 10.0.0.1 port 40578 ssh2: RSA SHA256:otVhNsBL6AYO2oq7y8I3VTqPa9of0q361Q7I052oURw Jan 24 11:45:43.864765 sshd-session[6316]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:45:43.881373 systemd-logind[1579]: New session 32 of user core. Jan 24 11:45:43.861000 audit[6316]: CRED_ACQ pid=6316 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:43.949916 kernel: audit: type=1101 audit(1769255143.857:946): pid=6316 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:43.950360 kernel: audit: type=1103 audit(1769255143.861:947): pid=6316 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:43.950426 kernel: audit: type=1006 audit(1769255143.861:948): pid=6316 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=32 res=1 Jan 24 11:45:43.974897 kernel: audit: type=1300 audit(1769255143.861:948): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe3a72c000 a2=3 a3=0 items=0 ppid=1 pid=6316 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=32 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:43.861000 audit[6316]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe3a72c000 a2=3 a3=0 items=0 ppid=1 pid=6316 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=32 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:45:43.861000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:44.041494 kernel: audit: type=1327 audit(1769255143.861:948): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:45:44.046741 systemd[1]: Started session-32.scope - Session 32 of User core. Jan 24 11:45:44.063000 audit[6316]: USER_START pid=6316 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:44.125691 kernel: audit: type=1105 audit(1769255144.063:949): pid=6316 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:44.073000 audit[6320]: CRED_ACQ pid=6320 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:44.167686 kernel: audit: type=1103 audit(1769255144.073:950): pid=6320 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:44.489827 sshd[6320]: Connection closed by 10.0.0.1 port 40578 Jan 24 11:45:44.491687 sshd-session[6316]: pam_unix(sshd:session): session closed for user core Jan 24 11:45:44.497000 audit[6316]: USER_END pid=6316 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:44.525494 systemd[1]: sshd@30-10.0.0.26:22-10.0.0.1:40578.service: Deactivated successfully. Jan 24 11:45:44.532565 systemd[1]: session-32.scope: Deactivated successfully. Jan 24 11:45:44.545867 systemd-logind[1579]: Session 32 logged out. Waiting for processes to exit. Jan 24 11:45:44.550404 systemd-logind[1579]: Removed session 32. Jan 24 11:45:44.552943 kernel: audit: type=1106 audit(1769255144.497:951): pid=6316 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:44.497000 audit[6316]: CRED_DISP pid=6316 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:44.594499 kernel: audit: type=1104 audit(1769255144.497:952): pid=6316 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:45:44.525000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@30-10.0.0.26:22-10.0.0.1:40578 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:45:44.673340 kubelet[2999]: E0124 11:45:44.671866 2999 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:45:44.678166 kubelet[2999]: E0124 11:45:44.676753 2999 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-7dbccbb54b-mvbw5" podUID="c101de33-34a1-459f-bcfc-2378435cf35e"